title
stringlengths 6
301
| keywords
listlengths 0
23
| messages
listlengths 2
2
|
|---|---|---|
Coping with enlargement: Regulatory reforms and third capital adequacy directive in the Polish banking sector
|
[
"Regulation",
"Banking",
"Poland",
"European Union"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper explores the idea of integration of Poland with the European Union on the example of banking regulatory reforms and capital adequacy directive (CAD3) implementation in the Polish banking sector. The aim of this work is to make a case for the implementation of the CAD3 rather than looking at the methods, modes and rules governing the implementation of CAD3.The paper is broadly divided into four parts. The first part covers a short introduction of the CAD3 followed by Poland's economic and financial transformation history; banking reforms specifically between 1982 and 1997. These can effectively be classed as preparation to enter the EU. The second part examines limitations of the Polish banking sector and the impact these limitations have on effective and immediate implementation and thus integration into the EU. These limitations are illustrated and followed by key socio-legal differences in banking regulatory practices and concepts in the European Union; specifically, the UK. Thirdly, the impact of Poland's accession on the limitations of the Polish banking sector will be presented with social banking as an example of that impact. Lastly, some of the legal and regulatory reforms in the Polish banking sector after EU enlargement will be discussed.\n2. Regulatory roots: economic and financial reforms: 2.1. International banking regulatory efforts\n3. Limitations of the Polish banking sector: key socio-legal differences in banking regulatory practices and concepts: 3.1 The integrated supervision proposal - Polish financial service authority powers\n4. Impact of accession on the limitations of the Polish banking sector: social banking example: 4.1 The issues related to social banking\n5. Legal and regulatory reforms in the Polish banking sector after EU enlargement: The Polish FSA model meets the demands of a growing banking and thus financial sector created by complex financial products and financial conglomerates.The Polish FSA structure proposal has been implemented on 19 September 2006 with the role of the Banking Supervision Commission (KNB) and its executive arm the GINB being taken over by the Polish FSA on 1 January 2008. It can be thus said that Poland now has the framework required to implement the necessary regulations.It is agreed that much depends on the institutional side of implementing CAD3 as is shown by the UK FSA example. Only a well \"built\" regulatory structure can administer the uncertainties of CAD3 with the correct regulation. Colloquially speaking having the right regulator structure is the first half of success. Implementing and administering regulations forms the other.Having chosen and established an institutional capital adequacy framework it is important for the Polish financial sector to establish a mechanism for implementing rules and regulations. The UK idea of the FSA handbook that can be tailored to each different credit institution can be utilised. It requires much work of the regulator but pay dividends when credit institutions are provided with the correct regulations swiftly.It is beneficial to compare a number of CAD provisions with their Polish equivalents to briefly exemplify the legal and regulatory reforms in the Polish banking sector after EU enlargement. The investigating of shareholders and board members of banks is very detailed in the EU, whereas in Poland less emphasis is placed on this important factor. The forms of start up capital that can be used in applying for a bank licence are unlimited in the EU, whereas Poland only allows personal cash assets of the licence applicants to be used. The rules for setting up foreign branches as well as rules on transfer of gain from those branches are sufficiently detailed in EU legislation. In Poland, such laws are at the project stage.European Union regulations as to the types of banking functions are cast much wider, and are inclusive of investment banking activity. Polish banking activity on the other hand still sees deposit-taking, loan making and merchant banking as the core bank functions to the exclusion of investment banking from distinct banking activity. This is a very important distinction. Poland distinguishes banks from investment firms and conglomerate institutions. That is to say banking is very much viewed as a standalone activity, (there is a structural divide between financial law and banking law) separate from dealing in shares and investing (historically undertaken by investment houses; \"domy maklerskie\"). Contrast this with the position of EU and Basel legislation which is tailored to multifaceted institutions drawing a distinction between credit institutions and investment firms only. An important area where Polish and EU capital adequacy standards are equivalent is credit risk assessment. It has been noted by commentators that this provision in particular is strictly a capital adequacy proviso. That is to say it is not a sufficiency provision, which would, in turn, cast a much more stringent requirement. It has been noted that more important issues like credit risk weighting have been brought in line with EU standards, while other issues including social banking are still to be addressed. Poland's membership in the European Union is still a challenge both for the newly established regulator and the banks functioning in Poland (Niewiadoma, 2008).In an emerging market such as Poland regulatory bodies should further have the power to require banks to hold higher level of capital than the minimal levels in CAD3. This is an example of gold plating. Such an approach is currently mainlined for start-up banks which are required to hold 15 per cent regulatory capital (instead of the required 8 per cent). This should be extended as a safe guard to be used on any bank that the regulator decides needs more regulatory capital. Care has to be taken however as \"gold plating\" regulations have been condemned by the banking industry in the UK for over regulating capital adequacy[22] (Financial Services Authority CP06/3, 5 May 2006).\n6. Conclusion: This paper explored the idea of integration of Poland with the European Union on the example of banking regulatory reforms and CAD3 implementation in the Polish Banking Sector. It made a case for the implementation of the CAD3 rather than looking at the methods, modes and rules governing the implementation of CAD3.The significance of choosing the right method for implementing the CAD3 has a number of implications. First and fundamentally, it will allow Poland, the biggest of the new EU member states, to show how it is capable of co-operating within the EU financial sphere by integrating with the EU financial market (e.g. joining the single currency union).Secondly, it is an important step to making the Polish financial sector internationally competitive. Thirdly, it allows the Polish banking sector a vital opportunity for growth pursuance. Fourthly, it will signal the major overhaul of capital adequacy laws and institutions in Poland since the fall of the Communism.The broad structure of the paper took into account a proposed mode of integrating Poland into the EU on the example of banking regulation. That is: understanding the historical transformation process that the state underwent; using international experiences in regulating the corresponding sector; and taking into account factors specific to emerging economies where the integration takes place. Examples of the last point are social banking as well as the need to create a regulatory body first and then to establish a hub from which regulations can be implemented and enforced.When implementing rules for capital adequacy the most important approach to keep in mind is the benefit of the banking sector and thus the regulated entities. The UK FSA displays what can be termed an institution friendly stance and this is the basis for the best implementation method of CAD3. An example of this is a waiver provision for the credit institutions that allows to bypass the more complicated supervisory requirements.The concept of \"copy out\" by which a literal copying of CAD3 provisions into FSA rules, it is maintained, is a very good idea that can ensure the transposition of the directive \"as it was designed\". At the same time the idea of \"gold plating\" could be introduced by increasing reserve capital requirements if needed.\nNotes: 1. J. Norton in Norton (1991) as well as Malgorzata Zaleska in Jaworski and Zawadzka (2005a, pp. 247-8). Despite the given name of Group of Ten the original committee is made up of the following countries: Belgium, Canada France, Germany, Italy, Japan, Luxemburg, Netherlands, UK, USA, Sweden and Switzerland.2. Another role of the Bazel Committee is the exchange of experiences as well as information regarding banking regulations.3. Past present and future.4. See Basel Committee Report.5. \"International Convergence on Capital Measurement and Capital Standards\".6. The 1999 standards are also spreading the ambit for Basle recommendations from credit risk to banking risk (operational, credit and market risks).7. Using value at risk techniques; mathematical formulae providing the amount that needs to be protected.8. Malgorzata Zaleska in Jaworski and Zawadzka (2005c, p. 255).9. This no doubt influences the purposeful approach of the EU banking regulation.10. Explenatory Note on the Law of Supervision of the Capital Market. (FSA-Model Proposal).11. Section 5 Law of Supervision of the Capital Market Explanatory Note.12. Section 7 Law of Supervision of the Capital Market Explanatory Note.13. Beginning of the process - it is important to note that another piece of legislation offering to amend the current pre FSA structure was issued as a \"rival\" idea/mode of implementing the CAD. This shall be discussed as an alternative and immediately below the FSA proposal.14. Opinion of the ECB of 9 March at the request of the Polish Minister of Finance on a draft law on the supervision of financial institutions. (CON/2006/15) p. 14.15. Opinion of the European Central Bank of 9 March 2006 on a draft law on the supervision of financial institutions (CON/2006/15), p. 14.16. See Jaworski and Zawadzka (2005d, p. 56) as sourced from Polish Central Bank Data.17. Opinion of the ECB of 9 March at the request of the Polish Minister of Finance on a draft law on the supervision of financial institutions. (CON/2006/15) p. 14.18. ECB Opinion on a draft law on supervision of the financial market and on amendments to certain laws. CON/2004/31 22 September 2004 Para 7.19. It is worth noting that the FSA in stating this puts the implementation of the Directive second to the regulation state of its market. By saying this it recognises the fact that the implementation is not an end in itself but merely a step towards a prosperous economy.20. Policy Statement on Strengthening Capital Standards 2, Feedback on CP06/3, Financial Services Authority, July 2006. Both available at: http://www.fsa.gov.uk/pages/Library/Policy/CP/2006/06_03.shtml21. Comments of the Reserve Bank of India on A New Capital Adequacy Framework para. 6.1 and 6.3.22. Joint Trade Associations' Response to the Financial Services Authority CP06/3: Strengthening Capital Standards, 5 May 2006.\n",
"role": "user"
},
{
"content": "Purpose: - This paper aims to broadly discuss the consequences of EU enlargement on the Polish banking sector and the issues arising from it.\nDesign/methodology/approach: - There are two broad objectives of the paper - first, to discuss the scope and evolution of banking and regulatory reforms since Poland's accession to the EU and second, to examine some of the specific legal and regulatory issues that have arisen from the implementation of third capital adequacy directive (CAD3) in the Polish banking sector.\nFindings: - Implications of CAD3 implementation in Poland include: Poland's capability of co-operating within the EU financial sphere; making the Polish financial sector internationally competitive; opportunity for growth for the Polish banking sector; and a major overhaul of capital adequacy laws and institutions. A proposed mode of CAD3 implementation as an example of integrating Poland into the EU include: understanding the historical transformation process that the state underwent; using international experiences in regulating the corresponding sector; and taking into account factors specific to emerging economies where the integration takes place. Examples include: social banking and the need to create a regulatory body first then implementing and enforcing regulations.\nOriginality/value: - This paper explores the integration of Poland with the European Union on the example of banking regulatory reforms and CAD3 implementation. It makes a case for the implementation of the CAD3 rather than looking at the methods, modes and rules governing the implementation of CAD3.\n",
"role": "assistant"
}
] |
Accounting and political parties: explaining the "why" of an Italian light touch regulation (1974)
|
[
"Political parties",
"Accounting",
"Legitimacy",
"Light touch regulation",
"Regulatory space",
"System of controls"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The multifaceted nature of accounting has been widely explored in the field of accounting history by analysing the role played by accounting and accountability systems in different contexts and in peculiar historical moments (Hopwood, 1983; Napier, 1989; Parker, 1997; Carmona and Zan, 2002; Carnegie and Napier, 2002; Carmona, 2004; Guthrie and Parker, 2006; Carnegie, 2014). Nevertheless, there seems to be a lack of academic interest in accounting regulation as it refers to political parties. Accounting and political parties, in fact, is an underexplored topic in the field of accounting history and this lack of interest is surprising, given the pivotal role played by parties from both a political and a social standpoint.\n2. Theoretical framework: 2.1 The regulatory space\n3. The system of controls introduced by the Law: The Law (specifically, articles 1, 2 and 3) introduced two forms of subsidies or contributions for political parties. The first entailed the assignment of a total amount of 15bn lire[3] as a contribution to the expenses that political parties bore for the re-election of the Parliament. The second entailed the assignment of a total annual amount of 45bn lire[4] as a contribution to the operating expenses of the various parliamentary groups[5] and of the related political parties. For every form of contribution, the Law specified which political parties were to receive it as well as the way the contributions were to be allocated among them.\n4. Outlining the regulatory space: 4.1 Historical timing\n5. Discussion: The analysis of the regulatory space has brought out that the Law was enacted in a context characterised by a struggle (Hancher and Moran, 1989; Larsson, 2005) between two cultural variables that were deeply rooted and in competition with each other. Both were potentially able to affect the output of the regulatory process.\n6. Conclusion: This paper aimed to explore why the Law established a system of controls inspired by light touch logics (Carr, 1997; Van de Velde, 1999; Howe and Landau, 2007). The empirical analysis has shown that the reasons for this action can be found in the peculiar configuration assumed by the regulatory space in which the Law was enacted (Hancher and Moran, 1989) and, in particular, in the need to find a trade-off between the different interests that emerged within the regulatory space.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explain the reasons why Law no. 195 of 2 May 1974, which established a system of public funding for the Italian political parties, introduced a system of controls that was light touch in nature.\nDesign/methodology/approach: The paper integrates the theoretical framework on regulatory space, proposed by Hancher and Moran (1989), with that of legitimacy (Suchman, 1995) to explain the peculiar nature of the system of controls introduced by the Law. Moreover, a set of primary and secondary sources is used to provide a full comprehension of the context, of the relationships among the actors involved in the regulatory process, and of the nature of the regulated issues.\nFindings: Besides showing that the nature of the output of a regulatory process can be understood as the effect of the peculiar configuration of the regulatory space in which it takes place, the study also sheds light on the role that legitimation can play with regard to the other features of the regulatory space, namely on its ability to strengthen or to limit their effects on the output of the regulatory process.\nOriginality/value: The paper deals with accounting and political parties which is a much underexplored topic in the field of accounting history. In addition, from a theoretical standpoint it contributes to extending the theoretical framework by Hancher and Moran (1989).\n",
"role": "assistant"
}
] |
Revisiting storey enclosure method for early estimation of structural building construction cost
|
[
"Regression analysis",
"Bootstrap",
"Range estimation",
"Construction cost",
"Storey enclosure method",
"Residential building projects"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Early predictions of likely project costs are fundamental to project evaluation, engineering design, cost budgeting and cost management. They play a vital role not only in construction project feasibility studies but also in construction projects' ultimate success (Cheng et al., 2009). In building construction, for key budgeting and feasibility decisions, construction cost estimates usually need to be predicted at early stages when project scope is not finalized (Sonmez, 2011). As design information available at early stages is still very coarse and limited (Cheung and Skitmore, 2006b), quantity take-off works cannot be performed to form detailed cost estimates (Sonmez, 2008). Conventionally, construction cost estimates are based on the experience of cost estimators and/or the database of unit rates from previous construction projects (Cheung and Skitmore, 2006b). In Vietnam, yearly documents of unit rates for many cities and provinces published by the Ministry of Construction (MOC) based on historical cost data are encouragingly used in cost estimating practices; however, their publication is usually very late. Although early cost estimates are not expected to be precise, inaccurate predictions could lead to lost opportunities, wasted development effort and lower than expected returns (Oberlender and Trost, 2001). In fact, high project performance and project success are not commonplace in the construction industry (Long et al., 2004). Especially, in developing countries, construction projects have encountered many problems which have been reducing project success (Le-Hoai et al., 2013). Recently, poor performance of construction projects has been blamed in Vietnam (Long et al., 2004). Cost overruns have arisen during project implementation (Le-Hoai et al., 2008). So far construction cost estimates have encountered difficult challenges in assessments of variability and confidence due to lack of reliable methods or supporting tools. In fact, the risks of related cost estimates are almost uncontrollable.\n2. Literature review: A literature study which focused on cost estimating methods was conducted. Unit method (UM), floor area method (FAM) and cube method (CM), which belong to a group of single rate models, are considered as conventional cost estimating methods. UM is employed when the estimator has enough historical data available from experience on a particular type of project to relate some end-product units to construction costs. This allows an estimate to be prepared for a similar project when the only major difference between the projects is their size. The use of UM is limited to public sector projects and/or to very early stages of project definition, where very little design has been undertaken (Akinsiku et al., 2011). FAM and CM, which involve the total floor area and the volume of a building, respectively, as their single variable (Cheung and Skitmore, 2006b), are the single rate methods widely used in preparing preliminary estimates. James (1954) developed an alternative single rate method, namely storey enclosure method (SEM), which was considered as the most sophisticated single rate method ever devised for early-design-stage cost estimates (Cheung and Skitmore, 2006a). SEM uses a weighting scheme to quantify the building characteristics (e.g. building shape, height and number of storeys) into aggregated storey enclosure \"units\" for pricing by a single appropriate storey enclosure unit rate (see Equation (1)). Despite a superiority against both CM and FAM (Skitmore, 1991), SEM was criticized for lack of confidence in the arbitrarily prescribed weightings (Ashworth, 1999) and lack of historical data available to estimate the suitable unit rates (Seeley, 1996). Generally, the subsequent development of multiple rate methods has overshadowed the interest in single rate estimating methods (Cheung and Skitmore, 2006a):(1)\n3. Research methodology: The existing parametric models (e.g. SEM and regression models) can quantify the impacts of building-related variables on the structural construction cost, but are deterministic. On the other hand, there is a lack of probabilistic approach which arguably reflects the uncertainty and variability included in the estimated structural construction cost better. Thus, this study uses nonparametric bootstrap method to establish range estimation for the structural construction cost. The processes and related techniques used in this study are presented in Figure 1. Main steps include: formation of base variables; identification of possible variables; model selection; and range estimation.\n4. Results of model development: 4.1 Explanation of regression models\n5. Range estimation: Based on the nonparametric bootstrap method, range estimation for five case projects was developed by the RPASEM model. In total, 200 bootstrap RPASEM models were developed using 200 bootstrap data sets (each data set contains 59 data points), which were randomly drawn with replacement from the original data of 59 projects. The cost estimates for five case projects were predicted using 200 developed bootstrap RPASEM models. Then, range estimation for five case projects was established based on the predicted cost estimates.\n6. Proposed cost estimating processes for application: For application, this study proposes two cost estimating processes, including cost modeling process (see Figure 2) and range estimating process (see Figure 3). According to Figure 2, in early stages of a building project, cost estimation models which predict point estimates are developed using historical data. As a result, a robust model is identified from the developed predictive models. This robust model will be used to establish range estimation according to Figure 3 as follows. After formation of design ideas (e.g. project scope, floor area and number of storeys), initial project designs are developed and design parameters are identified. Then, n cost estimates are predicted by the selected robust model, using n bootstrap data sets which are randomly drawn with replacement from the original sample of historical data. Finally, range estimation, including cumulative probability-based range estimates and confidence intervals, is established.\n7. Contributions: One of the main contributions of this study lies on range estimation using nonparametric bootstrap method. This could lead to the potential use of bootstrap method for range estimation in the construction industry. The second contribution lies on the use of cumulative probability-based range estimates and confidence intervals, which can provide information about the level of uncertainty and variability for the estimate, respectively. Such information can be very crucial for management decisions in early stages of any project.\n8. Limitations: While this study makes several contributions on the range estimation methodology, some limitations should be noted. One of the techniques used to develop range estimation is regression analysis, which requires a priori decision on the class of relations for variables (Sonmez, 2008). In addition, the bootstrap process of cost modeling for range estimation is used for residential building projects in Vietnam and, therefore, cannot automatically be employed for other types of building projects and/or in other countries without additional data collection. However, the proposed process can universally be applied to range estimation for project costs of any project types in any areas.\n9. Conclusions: This study developed various early cost estimation models for residential building projects using several preliminary design parameters. Regression analysis was used to develop the models. Input variables were formed based on modification of SEM and checked in terms of correlation and multicollinearity. An endeavor of applying principle component analysis to eliminate the multicollinearity was also conducted. The developed models were compared in terms of goodness of fit, bias, precision and also stability of prediction. An assumption of excluding the high-VIF variables may reduce the predictability of the regression models was verified. A robust model, namely RPASEM, was developed from an integration of ASEM, principle component analysis and regression analysis. The developed model can predict reasonably accurate cost estimates for residential building projects using design information available in early stages.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop several predictive models for estimating the structural construction cost and establish range estimation for the structural construction cost using design information available in early stages of residential building projects.\nDesign/methodology/approach: Information about residential building projects is collected based on project documents from construction companies with regard to the design parameters and the actual structural construction costs at completion. Storey enclosure method (SEM) is fundamental for determining the building design parameters, forming the potential variables and developing the cost estimation models using regression analysis. Nonparametric bootstrap method is used to establish range estimation for the structural construction cost.\nFindings: A model which is developed from an integration of advanced SEM, principle component analysis and regression analysis is robust in terms of predictability. In terms of range estimation, cumulative probability-based range estimates and confidence intervals are established. While cumulative probability-based range estimates provide information about the level of uncertainty included in the estimate, confidence intervals provide information about the variability of the estimate. Such information could be very crucial for management decisions in early stages of residential building projects.\nOriginality/value: This study could provide practitioners with a better understanding of the uncertainty and variability included in the cost estimate. Hence, they could make effective improvements on cost-related management approaches to enhance project cost performance.\n",
"role": "assistant"
}
] |
When do investors value board gender diversity?
|
[
"Gender",
"Board of directors",
"Event study"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The persistent challenge of achieving gender diversity in business is of great concern when examining corporate boards. Growth in the percentage of women on corporate boards has been slow. Female representation on corporate boards in countries such as Norway and France averages near 40 per cent; in others, such as the USA and Canada, it averages close to 20 per cent, whereas in Brazil, India and Portugal, the average is around 10 per cent (Eastman et al., 2016). Yet, even if boards achieve gender diversity, there is a lack of consensus as to whether investors believe gender diversity will create a board that is more able to execute its duties and increase firm value (Post and Byron, 2015; Kang et al., 2010).\nData and methodology: Background: Italian corporate governance and the Italian gender parity law\nResults: On the date of the Italian Parliament's passage of the gender diversity law, 45 per cent of Italian firms did not have a single female board member; 89 per cent of firms did not meet the law's initial requirement of at least 20 per cent female board representation; and, more than 98 per cent did not meet the law's eventual requirement of at least 33 per cent female board representation (Table I). Tables II and III contain additional data set summary statistics.\nDiscussion and conclusion: My paper contributes to two fields of research: corporate governance and gender diversity. Specifically, this paper examines the immediate financial effects of mandated gender diversity in publicly traded firms through immediate investor reaction. The results show that investors anticipate a higher level of future cash flows from the expectation that females will occupy close to or greater than a critical mass of board positions. Researchers have explored links between firm financial outcomes and gender diversity in corporate boards. However, they do not establish causality or isolate the perceived value of gender. This paper helps to bridge that gap by demonstrating that investors place a higher value on firms that will need to add more women to meet the Italian government's mandated gender diversity levels. In a one-day event study, investors reacted to the news of Italian government mandated percentage of female board representation, not the appointment of specific women, by increasing the value of firms that fell short of the mandated gender diversity requirements at that time. In other words, investors reacted to the concept of gender diversity without knowing the actual identities or qualifications of the women to be appointed.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to use the passage of the Italian Gender Diversity Law to help isolate the effects of board gender diversity on firm value by investigating conditions under which such diversity provides greater role-enhancing resources to the board.\nDesign/methodology/approach: This paper used a one-day event study to measure when gender diversity matters to investors. Abnormal returns from Italian firms were used to study investors' anticipated outcomes of the effect of gender diversity on firm value.\nFindings: Board gender diversity is financially beneficial especially for firms with a male dual CEO and board chair and with few or no women on board committees and firms that operate in industries with greater levels of competition. Addition of these moderators more than doubles the variance explained. Moreover, the effect of gender is isolated in this study, which examined investor reaction to the expectation of increases in the number of female board members, rather than to specific female appointees.\nSocial implications: Determining the conditions when a gender diverse matters to firm value is important for shareholders, policymakers and advocates for gender equality. The findings illustrate precise conditions for stakeholders to make the case for board gender diversity as achieving financial reward, in addition to societal benefit.\nOriginality/value: The value of a gender diverse board is contingent on the company's need for diverse resources (e.g. more competition, lack of gender diversity on committees or CEO duality). This paper provides insight as to why prior research linking board gender diversity to firm value finds seemingly contradictory results. Thus, this paper provides useful insights for researchers, boards and legislative bodies.\n",
"role": "assistant"
}
] |
How to combat abuse at work: Online programs to help educate the organization
|
[
"Respect",
"Incivility",
"Training programs",
"Workplace abuse"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nWorkplace abuse and its impact on employees: One of the more worrying aspects of business currently is the prevalence of hostile behavior in the workplace. Various studies report that a high percentage of employees have been subjected to some form of abuse over recent years.\nNegative consequences for the organization: Of course, the company hardly gets away scot-free either. When abuse rears its ugly head, serious repercussions are virtually inevitable. Employees affected will become more prone to:\nOnline training interventions and what they can achieve: Through training and education initiatives, employers can:\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: Online training interventions offer a means of tackling the growing and serious issue of abuse in the workplace. Such programs can be targeted at the whole organization so that everyone becomes more informed about different types of abusive behavior. On this evidence, education helps reduce the incivility that often progresses to more serious forms of abuse. Individuals can simultaneously learn to become more respectful toward others using programs that further appeal because of their flexibility.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Fashion and women's self-concept: a typology for self-fashioning using clothing
|
[
"Consumer behaviour",
"Consumer attitudes",
"Fashion leaders",
"Apparel",
"Clothing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In today's society, particular emphasis is placed on material items and the construction of one's image. Possessions, as stated by Belk (1988), have the ability to extend the self and can act as a symbol of self-definition. Clothing can be used as a tool of self-presentation where an individual can create, maintain and modify their image to align with perceived social norms. Prior research has established that an individual has the ability to identify themselves and others through clothing (Solomon and Schopler, 1982; Roach-Higgins and Eicher, 1992), but gaps remain in regard to how this takes place. Literature has also established that there are key motivational drivers behind the construction of one's identity and the particular role portrayed through this construction, but the belief that women will all engage with fashion in a similar, and personally consistent, way often clouds our understanding of the different fashion-related identities expressed by individuals. Some authors note a need for further research into the understanding of the ways in which different contexts alter self-presentation (Tice et al., 1995), which supports a need for examination of women with markedly different emotional attachments to fashion and fashion-related behaviours.\nLiterature: Self-concept theory\nMethodology: In-depth interviews\nResults and discussion: Internal impression management pressures\nA typology of fashion identity: When the emergent themes are compared across the women who participated in this study, five clear fashion identity types emerge from the data. Each of these fashion identities are discussed below, using a case approach to present the discussion in reference to five women who exemplify the typology.\nConclusions: A link between an individuals' fashion orientation (self-fashioning, following, rejection, etc.) and their self-concept has been discussed in prior research (e.g. Gutman and Mills, 1982; Goldsmith et al., 1999; Park and Burns 2005), but the common assumption is that those with high levels of self-esteem, or a strong self-concept, are more likely to exhibit self-confident, and independent fashion behaviours, whereas those with low self-esteem are likely to be fashion followers or non-engaged with fashion. This study sought to examine the clothing consumption experiences of a group of women, active in their engagement with fashion. The research finds that both security of self-concept, in relation to fashion and general self-esteem, as well as insecurity, can motivate women towards moderation of self via fashion choices. In fact, the relationship between an insecure self-concept and independent fashion behaviours (making \"risky\" and extreme clothing choices) seems stronger than that of the relationship between fashion independence and a strong self-concept.\n",
"role": "user"
},
{
"content": "Purpose: An individual's identity is defined in the role that they devise for themselves, based on social positions. Examining identity motives can help in understanding what influences one to take on a particular role. Self-esteem is one of the major motivational drivers in determining the role that an individual takes on. Individuals, through self-presentation, are said to be motivated to control the impressions others form of them. In this way, self-concept and fashion innovativeness are linked - with prior research suggesting that those with high levels of fashion innovativeness are also those with a strong sense of self. Where a gap remains, however, in exploring the direction of the relationship between self-concept and being more innovative and fashionable in clothing choices, as well as how individuals reflexively judge their own fashion choices against their perception of others - e.g. can you force yourself to be a fashion leader? The paper aims to discuss these issues.\nDesign/methodology/approach: This study takes a lived experience approach to examine fashion as a tool in establishing social hierarchies amongst women. The study uses depth interviews with ten women to explore the developed self-concept of women actively engaged with fashion consumption.\nFindings: The research presents a typology of fashion identities, exploring notions of security, dominance and innovativeness in self-fashioning using clothing.\nResearch limitations/implications: The research is exploratory, and limited to a sample of ten women. However, the study offers a number of key findings to drive future research in this area.\nPractical implications: The research finds that both security of self-concept, in relation to fashion and general self-esteem, as well as insecurity, can motivate women towards fashion independence. This suggests that identity-based marketing is likely to be more successful than lifestyle-based marketing, when selling women's fashion clothing.\nSocial implications: In prior research, self-concept and fashion innovativeness are linked - with prior research suggesting that those with high levels of fashion innovativeness are also those with a strong sense of self. This study finds that those with an insecure sense of self may also exhibit fashion independence, using fashion to acquire social capital.\nOriginality/value: This paper illustrates the concept that, unlike previous notions of fashion independence and engagement with fashion, these fashion-involved categorisations of behaviour are not always driven by sophistication, confidence, creativity and low fear of risk. Instead, this study has shown that fashion innovativeness can be motivated by an overarching fear of the outcomes of being judged unfashionable.\n",
"role": "assistant"
}
] |
Investigating the transformative impact of bank transparency on consumers' financial well-being
|
[
"Banking industry",
"Transformative service research",
"Financial services",
"Financial well-being",
"Bank transparency"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Financial well-being (FWB) has recently emerged as an important understudied topic. The Transformative Service Research (TSR) movement calls for research that centers on creating uplifting changes and improvement in the well-being of individuals and societies (Anderson and Ostrom, 2015; Rosenbaum, 2015). Inspired by this movement, the foundational paper of Bruggen et al. (2017) sets the scene for the overlooked concept of FWB. In this respect, FWB is defined \"as the perception of being able to sustain current and anticipated desired living standards and financial freedom\" (Bruggen et al., 2017, p. 229).\nLiterature review: Financial well-being\nMethodology: Data collection\nResults: Measurement assessment\nDiscussion: In light of the increasing pressure on banks to implement interventions improving the well-being of their customers, the purpose of this research was to explore if and how bank information transparency relates to FWB. This paper contributes to the debate on the impact of information transparency on consumers' FWB (Alsemgeest, 2015), and shows that bank information transparency has an uplifting effect on consumer well-being.\nLimitations and suggestions for future studies: This study has a number of limitations that suggest potential avenues for further research. This paper focuses on the cognitive side of attitude formation; however, as the literature suggests, attitude formation is also an affect based-information process. Thus, future research could evaluate the affective response of customers based on the affective evaluation of the degree of perceived disclosure, accuracy and clarity of shared information. In addition, future research could evaluate the interaction of the cognitive and affective evaluation of the shared information in shaping attitudes toward banks and the perceived financial self-efficacy of the customers.\n",
"role": "user"
},
{
"content": "Purpose: Grounded in Transformative Service Research, the purpose of this paper is to explore the mechanisms by which bank information transparency influences consumer's financial well-being (FWB). The authors propose that customer attitudes toward the brand and the subjectively perceived ability of individuals to deal with the financial challenges explain the enhancement of FWB driven by bank information transparency.\nDesign/methodology/approach: A survey was conducted to test the proposed hypotheses. In total, 400 bank customers of five commercial banks in Colombia were approached and asked to fill out a pen and paper questionnaire. Serial mediation analysis was applied to test the hypotheses.\nFindings: This research shows that bank information transparency can uplift the FWB of customers. Furthermore, the positive effect of bank information transparency on the FWB occurs because the shared information improves the positive attitudes toward banks and the perceived financial self-efficacy of customers.\nResearch limitations/implications: This paper heeds the call of current literature for improved explanations of the relationship between attempts to inform consumers about financial services and their FWB.\nPractical implications: This research shows that managers who embrace the challenging task of improving the FWB of their customers should design strategies for more transparent information sharing with their customers. However, these strategies should be designed not only to deliver information to customers but also to increase the perceived disclosure, accuracy and clarity of shared information.\nOriginality/value: This pioneering study aims to explain the effects of bank information transparency on the FWB of consumers by drawing on interdisciplinary literature. This research is important as many banks aim to increase their information transparency without a clear understanding of the effects of these actions on consumers and therefore in many instances their efforts fail. A key contribution of this study is identifying concrete mechanisms (i.e. brand attitudes and self-efficacy) that help managers to improve customers' FWB via information transparency. Accordingly, the authors offer suggestions for better information transparency strategy implementation.\n",
"role": "assistant"
}
] |
Perceptions and experiences of early-adopting registered dietitians in integrating nutrigenomics into practice
|
[
"Perceptions",
"Qualitative",
"Nutrigenomics",
"Personalized nutrition",
"Registered dietitians"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Nutritional genomics (NGx) is an emerging field focused on interactions between food, nutrition and genes (Ferguson et al., 2016). Increased understanding of gene-nutrient interactions may facilitate health and disease prevention (Casas et al., 2016; Corella et al., 2016; Celis-Morales et al., 2016). A growing market offers genetic tests \"direct-to-consumer (DTC)\" as well as via healthcare professionals. The Food4me project has illustrated how personalized nutrition can be delivered online to the public (Celis-Morales et al., 2016). Tests can deliver information linking diet to health, lifestyle, weight or improved fitness (Covolo et al., 2015; Bloss et al., 2011). Omics technologies (metabolomics, lipidomics and transcriptomics) enable highly personalized and targeted approaches to dietary health promotion (Sun and Hu, 2016) which have already been shown effective for outcomes related to cardiovascular disease (Fito et al., 2016). Consumer interest is high and demand for trained practitioners expected to increase (Berezowska et al., 2015). Registered dietitians (RDs) will be key professionals for translating the science of nutrigenomics into practice (Berezowska et al., 2015; Abrahams et al., 2017; Stewart-Knox et al., 2013, 2016; Fallaize et al., 2015). The Academy of Nutrition and Dietetics, however, do not consider the field ready for routine practice (Camp and Trujillo, 2014), so that application of nutritional genomics has tended to be low amongst the dietetics profession (Collins et al., 2013; Whelan et al., 2008).\n2. Method: 2.1 Sampling\n3. Results and discussion: Data were best described by four main themes (profile of the NGx practitioner, experiences of RDs in practice, perceived barriers to NGx and perspectives on the future and sub-themes (Figure 1). These provide insights into NG practice from the perspective of practicing RDs, and convey their opinions and views on how clinical practice and training could and should evolve, and how best to deliver such services in the future.\n4. Conclusions: This research has sought insight into the experiences and perspectives of RDs who have taken the leap and ventured into NGx practice. Early adopters of NGx were experienced, self-efficacious RDs who actively sought knowledge of NGx through communication at conferences and other media. By virtue of being in the private sector they had autonomy and were able to apply their new knowledge in practice. NGx was considered an extension of current practice for which RDs already had the skills. Perceived skepticism among peers about the efficacy of NGx was perceived to deter adoption of NGx and was blamed on the unregulated proliferation of websites offering tests and \"fad\" diets. Reluctance to adopt nutrigenomics among the wider dietetics community was also attributed to concern about the potential to widen health inequality by catering to the worried well to the neglect of sustainable public health models for delivery of services to the wider population (Fischer et al., 2016). None of the RDs practiced across national borders so worked along-side national healthcare provision which apparently varied in the level of service offered. Interviewees in the UK and Israel referred to how NGx testing seemed far removed from national healthcare provision, which focused on treatment of acute chronic diseases to the neglect of health promotion and prevention. In Australia and South Africa, in contrast, private medical insurance in some cases allows earned reward points to be used to purchase NGx tests so that no \"out of pocket\" spend is required. In the Canadian focus group, one participant mentioned how their healthcare system allowed individuals could buy a test privately and be later reimbursed for the dietetic consultation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the perceptions and experiences of early adopters of the technology.\nDesign/methodology/approach: Registered dietitians (RDs) (n=14) were recruited from the UK, Canada, South Africa, Australia, Mexico and Israel. Six qualitative interviews and two focus groups were conducted online using a conference calling platform. Data were recorded, transcribed and thematically analyzed.\nFindings: Early adopters of nutrigenomics (NGx) were experienced, self-efficacious RDs who actively sought knowledge of NGx through communication with one another and the broader scientific community. They considered NGx an extension of current practice and believed RDs had the skills to deliver it. Perceived barriers to widening the application of NGx were linked to skepticism among the wider dietetics community. Proliferation of unregulated websites offering tests and diets was considered \"pseudoscience\" and detrimental to dietetics fully embracing NGx. Lack of a sustainable public health model for the delivery of NGx was also perceived to hinder progress. Results are discussed with reference to \"diffusion of innovation theory.\"\nOriginality/value: The views of RDs who practice NGx have not been previously studied. These data highlight requirements for future dietetic training provision and more inclusive service delivery models. Regulation of NGx services and formal recognition by professional bodies is needed to address the research/practice translation gap. Further research is required to inquire as to the views of the wider dietetics profession.\n",
"role": "assistant"
}
] |
What "career success" means to blue-collar workers
|
[
"Careers",
"Career development",
"Career satisfaction",
"Blue collar workers",
"France"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The concept of a career has been the focus of studies by researchers, particularly researchers in management, for a long time. However, these studies have mostly related to specific populations, with executives and senior managers currently being privileged subjects. Although they represent nearly a quarter of the French working population, blue-collar workers are very seldom the object of career studies.\nA definition of career success: the perspective of blue-collar workers: Definition of career success\nMethod: Population and sample\nResults: Although every worker gave his/her own definition of success, many common criteria appeared even within the small sample used in this study. First of all, to create categories, the data were conceptualized. For example, the sentence, \"For me, success is to have a very high wage, to have money\" was encoded as monetary reward. Then, concepts which seemed appropriate to a similar phenomenon were regrouped together (Strauss and Corbin, 1990). Concepts of monetary rewards and fringe benefits referred to the same phenomenon: evaluation of success in terms of material criteria. During the codification procedure, they were regrouped in the category called \"material career success\". Once codification was complete, 11 facets were regrouped in three categories of success: material, psychological and social. If success traditionally comprises an objective and a subjective part, it is interesting to note here that the perception of success was influenced by a material aspect (rather close to objective career success) but also by psychological and social aspects, the latter often being absent from research (although the study of Lau (2002) on business owners did adopt this dimension).The three categories are represented and described below (Figure 1).Criteria for material career success\nDiscussion: The workers interviewed have mostly assimilated the fact that their careers do not provide upward mobility. So, their expectations do not rest on high-hierarchical position but more on psychological aspects. They try to compensate the lack of upward mobility by task enrichment, continuous training and status recognition. Studies with other populations found upward development was important, even though it is difficult. For example, the study of Megemont and Le Blanc (2006) on 251 technicians showed the link that still exists between upward mobility and self-esteem for those workers. The study by Falcoz (2001) on executives also showed the important role of traditional hierarchical progress for them. It is not about saying that blue-collar workers do not wish upward mobility but that this is rare. So, they develop other aspects to define career success.Concerning material success, blue-collar workers use criteria close to other populations, the more important remaining monetary recompense (Nicholson, 2000). While research may often imply that the workers have an entirely instrumental vision of work, important psychological aspects emerged in the interviews. Even though many studies deny that these occupations have autonomy or challenge (Lucas, 2002), blue-collar workers frequently cited accomplishment and job success. Even if they do not have these, they seek them to be able to have feelings of success.For material and psychological measures, gender has an influence on the relative importance of different criteria. This difference is similar to the differences observed in other samples (Kirchmeyer, 1998).Many criteria used by blue-collar workers are those that have been found to be used by executives in other studies (Sturges, 1999) like wages or recognition. Finegold and Mohrman (2001) underlined the importance of life-balance. This dimension appeared for blue-collar workers too. However, the relative importance of these criteria seems to differ between the two populations. For example, autonomy is mentioned by blue-collar workers but it belongs to the facet of job success, giving it a different perspective on work overall. Further studies would be necessary in order to compare perceptions of career success more precisely between executives and blue-collar workers.It is important to stress the importance of social results, proving the importance of position in society and the opinion of others in individual careers (Heslin, 2003). Few previous studies have highlighted the importance of this dimension (Lau, 2002). The development of this aspect shows the importance of studying the standards of employees but also shows that other expectations influence perceived career success (Heslin, 2003).The present study illustrates the fact that blue-collar workers seek different elements in their work than those which are identified by their organisation (i.e. money), as Cangemi and Guttschalk (1986) have already demonstrated. That shows the need to let blue-collar workers explain themselves in order to understand their career expectations. More generally, this underlines the importance of studying career success in different populations because of systematic variations between professions (Heslin, 2005).It is also necessary to improve the measurement of subjective career success by studying different facets, some perhaps still unknown, and developing scales adapted to the values, careers and expectations of different groups. Some studies in different national contexts should be considered too, in order to see how cultural differences influence work evaluation (Triandis, 1994).Individuals are different in the way they conceptualise their career success. For firms, it is essential to understand better the expectations of their employees, whatever their hierarchical level. They can use this information to attract, motivate and retain their staff. However, blue-collar workers often work in organisations that pay little attention to their conception of career and career success. In a changing economic environment, with an increasingly qualified workforce, firms must try harder to consider the expectations of their employees in order to avoid problems, such as excessive turn-over.\n",
"role": "user"
},
{
"content": "Purpose: - Career success such as it is classically modelled and measured is not necessarily representative of the perceptions of blue-collar workers. Therefore, the goal of the present study is to understand what it means to succeed in blue-collar occupations. A definition and a measure are useful starting points in explaining the individual's evaluation of success and in analysing the behaviour which results from it.\nDesign/methodology/approach: - After a review of literature, in order to develop the theoretical framework for the research, 25 exploratory interviews were conducted with blue-collar workers in order to obtain their criteria for career success. The results are reported, together with the methodological decisions and a proposed explanatory model.\nFindings: - Success is traditionally described as having an objective and a subjective part. In order to understand the perceptions of blue-collar workers, it is necessary to recognise that their perceptions are influenced by material, psychological and social aspects.\nPractical implications: - The definitions developed in this paper could be used to develop employment programmes addressing the expectations of blue-collar workers, and to attract new employees to this kind of occupation. More generally, with the flattening of organizations, it is necessary to re-evaluate the concept of success for all workers, and to develop policies that are appropriate to changes in the labour market.\nOriginality/value: - The majority of career studies focus on \"white collar\" workers. Traditional career theory has developed models which consider that all the employees are guided by the same systems of values (status, power, wages, etc.). Thus, this paper fulfils a need to develop an understanding of career success from the perspective of blue-collar workers.\n",
"role": "assistant"
}
] |
University brand image as competitive advantage: a two-country study
|
[
"Heritage",
"Service quality",
"Higher education",
"Reputation",
"Brand image"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Competition in the higher education sector is intensifying (Berry and Cassidy, 2013). In the USA, a number of universities have been consistently rising as barriers to entry continue to decline. Universities are increasingly resembling one another and are finding it difficult to attract quality students due to lack of a differentiating factor. Off late, the bargaining power has tipped toward students as they have more options to select from. Thus, the competitive advantage of universities is eroding.\nTheoretical framework: The higher education sector can be predominantly considered as a service industry (Durvasula et al., 2011). Educational services are intangible, heterogeneous, inseparable from those who deliver it, perishable and variable. University services can be classified into search, experience and to an extent credence attributes (Darby and Karni, 1973; Nelson, 1970). Search attributes are those that can be verified prior to purchase through readily available information source or direct inspection while experience attributes can only be verified upon its usage. Prospective students search for various attributes to gauge the quality of the university - its ranking, quality of faculty, infrastructure and campus visits to determine if the university satisfies their preference. However, it is not possible for the student to gauge the true quality of the university until he is enrolled and immersed in the university experience. Before enrollment, the student had evaluated tangible information that was readily available. However, he is only truly able to determine the service quality after enrollment (i.e. faculty teaching, quality of support, maintenance of infrastructure, etc.) - experience attribute.\nHypotheses development: Existing literature has cited multiple ways to increase satisfaction level among students. Brodie et al. (2009) include factors such as brand image, company image, employee trust and company trust to influence brand perception among customers. These brand perceptions affect the consumer's perception of reputation and, hence, their level of satisfaction. Casidy (2013, 2014) focus on the role of service quality and product brand orientation on student satisfaction.\nResearch method: Research design\nResults: Regression analysis was used to test the impact of university heritage, trustworthiness, service quality and reputation on student satisfaction in the American and Indian context. Findings support the proposed hypotheses in both the American and Indian contexts. All variance inflationary factor values of independent variables in this model are less than 2, indicating that no significant multicollinearity problem exists.\nDiscussion: Findings indicate that a university's brand image plays an important role in creating a distinct image in the minds of its students resulting in its competitive advantage. Positive brand image results in increased reputation and is positively associated with the satisfaction level of students, which would eventually lead to tangible and intangible benefits for the university.\nConclusions: This study outlines the association between a positive brand image and reputation on the level of student satisfaction. The study is important as it offers a road map for universities to distinguish themselves from their peers and develop a sustainable competitive advantage. The study offers three key aspects that universities need to work on - building their heritage, delivering great service quality and building trust among their key stakeholders - students. Having achieved their goal of achieving a positive perception, universities should relentlessly work on living up to it, so that their consumers - students - are satisfied. Satisfied students are one of the key sources that generate positive brand equity for the university, give back and recommend others to their alma mater.\n",
"role": "user"
},
{
"content": "Purpose: Given the competitive landscape in the higher education setting, it is important that universities adopt strategies that create competitive advantage for them. Universities must leverage their resources efficiently to address this goal. Creating a positive brand image is one such strategy. The purpose of this paper is to conceptualize university brand image as its heritage, service quality and trustworthiness and investigate their relationship with student's satisfaction. It also investigates the role of university reputation as a mediating variable.\nDesign/methodology/approach: Data were collected through a mixed method approach. The first stage involved qualitative interviews and focused group discussions with students to understand the factors responsible for student satisfaction with their respective universities. The second stage involved administering a survey questionnaire in two geographies - the USA and India to investigate the hypothesized relationship. The authors use regression analyses to test these relationships.\nFindings: Findings indicate that a distinct brand image plays an important role in students' level of satisfaction across both the USA and India. Service quality has a greater impact on student satisfaction levels across both contexts (as compared to university heritage and trustworthiness). The authors also find a positive mediating effect of university reputation in the relationship between university brand image and student satisfaction levels.\nOriginality/value: The current research contributes to the services marketing literature in the university context. It offers a framework for decision making in universities. It suggests that universities must work toward developing their brand image by focusing on its three dimensions - heritage, trustworthiness and service quality.\n",
"role": "assistant"
}
] |
Job demands-resources and employee health and well-being: The moderating role of contract type
|
[
"Job satisfaction",
"General health",
"JDR model",
"Multi-occupational sample",
"Multinational sample",
"Type of contract"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: For somewhat more than a decade, researchers have already published tens of studies in which the Job Demands-Resources (JDR) model serves as the main theoretical framework (Demerouti and Bakker, 2011). Many of these studies have provided ample evidence for the main effects of job demands and job resources on employee health and well-being, and considerable evidence for the buffer effect of job resources on the relationship between job demands and health and well-being (Bakker and Demerouti, 2007; Demerouti and Bakker, 2011). However, only few studies have expanded the JDR model by including moderating variables. This is remarkable, because it is quite unlikely that the basic premises of the JDR model pertain equally to all individuals. Because variations may occur in the extent to which job demands and job resources impact employees' health and well-being, Bakker and Demerouti called researchers to study the impact of different kinds of moderators on the JDR model (e.g. Bakker and Demerouti, 2007; Demerouti and Bakker, 2011). Several researchers have answered this call by testing the moderating role of personal characteristics such as self-efficacy, work orientation, and regulatory focus (Brenninkmeijer et al., 2010; Demerouti et al., 2012; Xanthopoulou et al., 2007). Studies that have expanded the JDR model by other variables than personal characteristics seem to be scarce. A study by Dollard and Bakker (2010) that examined the role of psychosocial safety climate as a precursor and moderator of the JDR model is a noteworthy exception. The purpose of this study was to examine the impact of another unique moderator on the JDR model, namely type of contract (i.e. temporary work vs permanent work).\nMethod: Sample and participant selection\nResults: Inspection of Table I reveals that the correlations between job demands, job resources, and job outcomes are in the expected direction. General health was negatively related to job insecurity (r=-0.14, p<0.001) and time pressure (r=-0.06, p<0.001) and positively related to autonomy (r=0.12, p<0.001) and social support (r=0.17, p<0.001). In addition, job satisfaction was negatively related to job insecurity (r=-0.23, p<0.001) and positively related to autonomy (r=0.42, p<0.001) and social support (r=0.44, p<0.001). The only correlation that was not in the expected direction was the correlation between time pressure and job satisfaction. This relation was expected to be negative, but turned out to be positive (r=0.04, p<0.05). Finally, Table I showed that permanent workers reported poorer job satisfaction (r=-0.06, p<0.001) and general health (r=-0.07, p<0.001) than temporary workers. This is in line with our predictions.\nDiscussion: In this paper, it was tested whether the main premises of the JDR model (i.e. the main effect of job demands, the main effect of job resources, and the buffer effect of job resources) are moderated by type of contract (i.e. temporary contract vs permanent contract). Hypotheses were tested by multilevel analyses. Contrary to previous research on the JDR model (Bakker and Demerouti, 2007; Demerouti and Bakker, 2011), results showed only moderate support for the main effects of job demands (job insecurity and time pressure) and job resources (autonomy and social support from the supervisor) in the prediction of job satisfaction and general health. We found no significant effects for time pressure and also the main effect of autonomy on general health appeared to be non-significant. Further, the evidence for the buffer effect of job resources turned out to be weak. As predicted, autonomy buffered the negative relation between the two types of job demands and job satisfaction. However, for social support we only found two reversed buffer effects. With respect to the moderating role of type of contract, evidence was also weak. Of all interactions tested, only three interactions were in line with our hypotheses. We found strong support for the two-way interaction between job insecurity and type of contract and moderate support for the three-way interaction between time pressure, autonomy and type of contract. Overall, our study showed more significant results for job satisfaction than for general health.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this paper is to investigate whether the main propositions of the job demands-resources (JDR) model are moderated by type of contract (i.e. temporary contract vs permanent contract).\nDesign/methodology/approach: Survey data were collected in a large, heterogeneous sample from different countries, sectors, and jobs (n=3,845). Hypotheses were tested by means of multilevel analyses.\nFindings: Results showed moderate support for the main effects of job demands (job insecurity and time pressure) and job resources (autonomy and social support) and weak support for the buffer effect of job resources in the prediction of job satisfaction and general health. The impact of contract type on the main propositions of the JDR model appeared to be weak. Yet, the evidence that was found suggests that temporary workers may be more tolerant to job insecurity and more likely to benefit from the buffering role of autonomy than permanent workers.\nOriginality/value: This is the first study to investigate whether the relation between job demands and job resources and employee health and well-being differs for permanent workers and temporary workers.\n",
"role": "assistant"
}
] |
Proposing micro-macro HRM strategies to overcome challenges of workforce diversity and deviance in ASEAN
|
[
"Diversity",
"Globalization",
"ASEAN",
"Workplace deviance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Due to globalization, member countries of Association of Southeast Asian Nations (ASEAN) have attained a notable economic growth and emerged as major manufacturing centers. If ASEAN were a sole unit, it would rank as the seventh largest economy in the world, behind the USA, China, Japan, Germany, France, and the UK (The ASEAN secretariat, 2013). To turn into a provincial manufacturing center driven by skills, innovation, and ingenuity, ASEAN will require more related tutelage and professional training for young women and men (Yap, 2014). But, current trends indicate a tremendous change in demographic dividend, which can be represented by a huge disparity in wages and gender representation among ASEAN members. In terms of salary, the largest discrepancies in salary were approximately 45.5 percent among middle- and lower-level employees and 43.9 percent among executive directors in firms with a market capitalization of over $1 billion (Rakyan and Gentili, 2016; Goh, 2017). With respect to the ASEAN nations, there is a huge disparity in wages that ranges from US$119 per month in Laos to $3,547 in Singapore. Regarding the gender representation, while the participation of women in workplace between 2010 and 2013 was around 59 percent, that for men was about 82 percent. According to the latest \"Women on boards Hong Kong report,\" the representation of female directors has increased considerably in Malaysia, Hong Kong, and the Philippines (Goh, 2017). In fact, in a recent \"Global gender gap report,\" the Philippines ranked seventh among 144 countries assessed around the world for female representation at workplace (Tan, 2016). But, instead of higher enrolment of women tertiary education across the ASEAN region, the portion of women in the workforce is consistently lower than that of men. Additionally, \"2016 Global talent competitiveness index (GTCI) report\" revealed that women and lower skilled employees are often relegated to informal work with no legal protection or employment benefits (Maxwell, 2016). For example, in Vietnam, only 31 percent of working women are employed formally, while 69 percent are engaged in informal work. Similarly, in Singapore, not only do women earn less than men, but one in every ten is unable to gain formal work (Tan, 2016). And, it is assumed that by 2025, there will be a net increase of 14 million highly skilled jobs in ASEAN economy, which could stay vacant or will be filled by inexperienced and unskilled workers. Even unskilled employees who will manage to find employment, their major distress will be poor job quality (Yap, 2014).\nASEAN context: Founded in 1967, ASEAN today encompasses Brunei, Cambodia, Indonesia, Laos, Malaysia, Myanmar, the Philippines, Singapore, Thailand, and Vietnam. The establishment of ASEAN as one of the most highly integrated regional organizations is the reflection of the process of globalization. On the overall globalization level index, Ireland tops the list with the highest globalization level at 91.30. Among ASEAN nations, Singapore secured 5th position with the globalization index at 87.49, Malaysia ranked 26th, Thailand is at 41st, followed by Indonesia at 86th, the Philippines at 88th, and Cambodia and Vietnam each at 119th and 120th. Based on the above ranking on globalization level index, it can be deduced that there is a huge disparity among ASEAN nations in terms of phases of development (Suci et al., 2015). But in spite of being in extremely distinctive phases of development, ASEAN nations share massive growth potential and are one of the fastest-growing consumer markets in the world. ASEAN markets of Indonesia, Vietnam, and Myanmar are evolving as the future markets for multinationals and ASEAN-5, i.e. Indonesia, Malaysia, the Philippines, Singapore, and Thailand in 2013 attracted more foreign direct investment as compared to China (Vinayak et al., 2014). Since the late 1970s, ASEAN has outpaced the rest of the world on growth in gross domestic product (GDP) per capita and the intraregional trade. Moreover, ASEAN, if looked as a single country, will be among the top ten economies in terms of youngest population, with an average age of 27 years old. An Ernst & Young study estimated that two-third of the world's middle class will be from Asia by 2030 (Suci et al., 2015). After China and India, ASEAN has the third largest workforce in the world. And, it is important to note that economies are about people, as they are the engines for growth.\nLiterature: Globalization\nMethodology: In carrying out a systematic review, four steps were followed: first, search terms such as \"globalization,\" \"workforce diversity,\" and \"workplace deviance\" were used to gather information for developing this conceptual paper. Second, research on four databases pertaining to management, sociology, and economics was examined. The databases were Google scholar, EBSCO, ProQuest, and SCOPUS. Third, the time frame for filtering the papers was set between 1992 and 2017 and the papers before 1992 were excluded. Moreover, the papers and keywords were searched in English pertaining to the ASEAN context. Finally, it was ensured that the resulting articles are representative by repeating the filtering process (Adolphus, 2009).\nConceptual framework: The review of literature and the amount of paper itself indicated dearth of literature associating these three variables. Thus, to overcome this gap, this conceptual paper studies the impact of globalization on workforce diversity and workplace deviance in the ASEAN context (Figure 2).\nFindings: Strategies to overcome challenges of workforce diversity and deviance\nMicro HRM strategies: Micro HRM strategies refer to views on interaction between employees and employers of an organization. The three major parts of micro HRM perspective include organizational culture, HR practices, and the role of top management and managers. The role of top management and managers is to develop and communicate a strong, high-performing, and positive atmosphere where employees have clear understanding of their roles and they collaborate effectively in alignment with the organizational goals and vision.\nMacro HRM strategies: Macro HRM strategies refer to ways in which an organization interacts with its customers, competitors, local, and global environment. Successful business owners maintain a balance with all these forces to develop strong relationships and to contribute positively toward its customers, employees, stakeholders, and community. Macro HRM strategies ensure a positive high-performing workplace, delivery of superior products and service to customers, and competitive advantage.\nDiscussion: This study proposes that in addition to the positive impact of globalization on ASEAN economy, it also causes poverty, injustice, disparity in income, and loss of protection of minority groups, thus worsening the working conditions (Jayasuriya, 2008; Kilic, 2015). According to Roberson and Stevens (2006), discrimination at workplace is significantly related to organizational justice, which can be defined as the extent to which employees perceive workplace procedures, interactions, and outcomes to be non-discriminatory in nature. The three dimensions of organizational justice are distributive justice, procedural justice, and interactional justice. Distributive justice is the perceived objectivity of outcomes, procedural justice is the fairness of a company's policies and procedures, and interactional justice refers to open interpersonal relationships and transparency of information at all levels of workplace (Nojani et al., 2012). Organizational justice plays a vital role in forming the attitudes of employees, such that employees who are ill treated at workplace become dissatisfied and disengaged from their job and are more likely to deviate from the organizational norms (Furnham and Siegel, 2011). On the contrary, fairness in workplace policies leads to higher levels of job satisfaction and engagement among employees, which results in exhibition of positive behavior (Jafari and Bidarian, 2012). Thus, workforce diversity, organizational justice, job satisfaction, work engagement, and workplace deviance are very much linked with each other, such that when employees perceive fairness in organizational HRM practices of workforce diversity, they are more satisfied and engaged in their work, and thus exhibit higher constructive deviance and reduced destructive deviance (Ladebo, 2008). Therefore, organizations that fail to embrace diversity, eradicate discrimination, and injustice have an adverse effect on both its employees and customers (Kundu, 2001) (Figure 4).\nManagerial implications: The findings of this study will benefit the ASEAN organizations operating in global economy, which have mandated the organizations to work with a diverse workforce. Diverse workforce in addition to offering various benefits like unleashing creativity, innovation, and improved group problem solving is also associated with various challenges like poverty, injustice, disparity in income, and loss of protection of minority groups, which subsequently leads to workplace deviance. This research represents the first effort to establish an association between globalization, workforce diversity, and workplace deviance in the ASEAN context. This study proposes that in order to reap the benefits of globalization, ASEAN organizations require a sustained, systemic approach, and enduring commitment toward its diverse workforce in a form of micro-macro HRM strategies. Managers must ensure a bi-directional implementation of the suggested micro-macro HRM strategies for effective outcomes. Moreover, there should be a harmonization between both these strategies such that a change in macro factor must facilitate a change in micro HRM practices.\nFuture implications: Besides the organizational initiatives, government organizations, civil society, academic bodies, and international communities must continuously work toward the upliftment of the diverse and deprived workforce in ASEAN. Regulatory mechanisms and legal frameworks are also necessary to eliminate discrimination on the grounds of workforce diversity. Enterprises, institutions, and government should take appropriate steps to ensure that employees are aware of their obligations and rights stemming from fundamental rights, equal employment laws, and related legal requirements. Moreover, government policy frameworks must identify which industrial sectors are likely to develop over the medium and longer term and should work proactively to ensure the required skill development and employment for diversified workforce. For example, according to \"Robert Walters' annual global salary survey, 2016,\" the growing hotspots include e-commerce and digital modern technology, technical healthcare, insurance, and pharmaceuticals (Froehlich, 2016). In addition, the role of business community, education and training providers, and employers is crucial to keep the organizations abreast of changing technologies and practices in the workplace. Furthermore, the outside agencies, including civil society and international bodies must provide awards and incentives to the organizations promoting involvement of minority and disadvantaged groups. And most important, the members of ASEAN should develop practical projects to assist the newer members with human developmental research, as well as establishing programs to facilitate their speedier integration into ASEAN in order to benefit the ASEAN economy.\nConclusion: This study proposed that organizations with inclusive policies for the diverse workforce including mix of nationalities, age groups, gender, cultural backgrounds, and low skilled/inexperienced employees can attain competitive advantage in the long run as compared to the organizations that ignore these policies. The suggested policies not only have significant impact on the outcomes of an organization, but it also plays a crucial role in increasing the GDP of an economy. According to ADB (2016), a 1 percentage point increase in the income share of the top 20 percent is associated with 0.08 percentage point lower GDP growth, while the same percentage increase in the income share of the bottom 20 percent is related to 0.38 percentage point higher growth. Thus, to attain this growth in terms of both organizational and national output, the ten ASEAN countries must undertake the development of disadvantaged groups as a national, organizational, societal, and cultural imperative. And, this could be attained by a strong collaboration between the private and public organizations, academic institutions, government bodies, civil society, technical upfronts, and international communities.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to propose a conceptual framework associating globalization, workforce diversity, and deviance and suggest micro-macro HRM strategies to overcome challenges associated with the workforce diversity and workplace deviance.\nDesign/methodology/approach: A systematic review of literature of past 25 years was carried out with the key word \"globalization, workforce diversity, and deviance\" from several electronic databases.\nFindings: Findings propose micro-macro HRM strategies to be adopted by HR practitioners in Association of Southeast Asian Nations (ASEAN) organizations to manage workforce diversity and deviance in the age of globalization.\nResearch limitations/implications: The challenges due to workforce diversity may get worse because ASEAN is more incongruent in terms of phases of economic, social, cultural, and political advancement. Therefore, proposed model can be tested and compared in different ASEAN organizations.\nOriginality/value: There is a dearth of literature associating globalization, workforce diversity, and deviance. This paper bridges this gap by proposing a conceptual framework in the ASEAN context and suggests micro-macro HRM strategies to be adopted by HRM practitioners to overcome associated challenges with workforce diversity and deviance.\n",
"role": "assistant"
}
] |
Beyond satisfaction: customer attachment to retail banks
|
[
"Service industries",
"Customer relations",
"Banking",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Recent evidence shows that consumers may switch some or all of their business to another supplier even when they fully satisfied (Buttle et al., 2002). Satisfaction, it appears, is not enough to ensure a long-term relationship. A managerial imperative is to identify whether there are alternative or enhanced approaches to improving customer tenure. Our research contributes to that body of knowledge.Our research was conducted in the context of the UK retail banking market for several reasons. First, banks had been reporting a high-level of customer dissatisfaction, yet also experiencing a high-level of customer retention. Increasing competitiveness in the sector meant that switching costs were falling as banks made it easier for customers to shift their accounts to other suppliers (Ahmad and Buttle, 2002). Secondly, banks were interested in learning from our research because they felt it would enhance their prospects of retaining customers in the face of reduced exit barriers. The magnitude of the problem is indicated by evidence from market research organisation, Mintel (2004), which reported that at any one time some 5 million UK bank customers are thinking of changing providers. Thirdly, retail banking is a service of which the vast majority of British adults have experience. Mintel (2004) reports that the UK's population of 60 million people operates 72 million current accounts. Banking is therefore a subject on which a preponderance of people would be able to express informed opinion (Lovelock, 1983).\nLiterature review: Scholars have developed two major bodies of knowledge about relationships between people and organisations. We draw on these \"parent\" disciplines to develop our conceptualization and operational measures of customer-bank attachment. Researchers have focused on:1. Employee commitment to, or attachment to, organisations. The literature is located in the discipline of organisational psychology.2. Customer involvement in products or services. The literature is located in the discipline of consumer behaviour.Organisational psychology\nThe banking context: Although our focus is clearly upon customer attachment to banking service providers we have been unable to locate any empirical research that is firmly grounded in these parent disciplines. We acknowledge that there have been a number of published papers in recent years that have attempted to deliver an understanding of customer-bank attachment (Schneider et al., 1980; Schneider and Bowen, 1985; Paradise-Tornow, 1991; Parasuraman et al., 1992; Brown and Mitchell, 1993; Bloemer et al., 1998; Iniesta and Sanchez, 2003). Many of these papers are embedded in the organizational climate, service quality and customer satisfaction literatures and do not ground themselves firmly in the two parent disciplines described above. Our aim is to \"go back to basics\" in developing our concepts and measures. We have therefore chosen to conduct our research drawing heavily on these foundation disciplines.\nQualitative research: A total of seven focus groups and 39 one-on-one interviews were undertaken. All were audio or video recorded. Transcriptions were produced and data were analysed within NUD*IST. Comparison of the independent content analysis of two coders resulted in the identification of 1,033 attachment-related items. These revealed six major themes within the data. Three were based strongly in values: organisational, instrumental and relational values. The other three themes were familiarity, inertia and alienation.Organisational, instrumental and relational values attachment\nQuantitative research: We conducted three waves of empirical research, analysed 975 questionnaires, and ultimately developed two scales: one designed to measure the three forms of positive attachment, and the other measuring the other three forms of non-positive attachment. Our process followed the recommendations of Churchill (1979) for improving the quality of marketing constructs.\nPurification process: Stage 1\nExplained variance: The cumulative variance exceeded 70 per cent for the samples of Banks 1, 2 and 4 while it was 66 per cent for Bank 3 and 62 per cent for the Stage One sample. The cumulative variance for the overall combined sample was reasonably high (71 per cent). The factor loadings indicate that factor 1 (Organisational Credibility, OC) is the strongest factor among the three factors across all samples (accounting for 52 per cent of variance) followed by the RV factor (accounting for an additional 10 per cent) and the Value Congruency factor (an additional 9 per cent).Reliability of the positive attachment scale\nManagerial implications: Our research shows that customers can, and do, develop positive attachment to their banks. The positive attachment inventory (PAI) we have developed shows that attachment is grounded on three forms of value: faith in organisational competence, mutually aligned and congruent values, and positive social bonds that deliver high-levels of relational value. Whilst we are not yet in a position to claim that high positive attachment scores correlate with higher levels of retention, account balances and customer profitability, we have found a high-correlation with perceived service quality, satisfaction and intention to recommend the bank. These may be regarded as surrogate indicators of customer financial value (Dean, 2004). As switching costs fall for customers wishing to change service providers, and as differentiation between banks is enhanced, banks will be looking to bond customers to them for the longer term. As noted in the introduction, customer satisfaction scores are not a strong predictor of future bank customer behaviour. The PAI offers a theoretically sound, richer, multi-dimensional alternative to satisfaction scores. Banks can employ attachment scores in lieu of satisfaction scores, to create and test new models linking attachment to customer value. The items within the attachment scale give clear guidance to banks wishing to enhance customer attachment profiles. The OC dimension of attachment incorporates measures of customer confidence, trust, respect as well as the bank's perceived ability to deliver appropriate services that present value for money. The RV dimension is focussed on the delivery of social benefits to customers in interactions with bank employees. The VC dimension assesses the alignment between the personal values of the customer and the policies and practices of the bank. Administration of the PAI will indicate to banks how strongly or weakly customers are attached to the bank. It will reveal relative strengths and weaknesses in the attachment profiles of customers, indicating which customers might be at risk. For example, it may reveal that within the customer base there are segments that are particularly susceptible to deficiencies in OC. Similarly, the PAI can be applied to competitors' customers to identify valued individuals or segments that might be switchable.\nConclusion: We have explored the nature of customer-bank attachment starting from first principles. From a mix of qualitative and quantitative analysis, we have generated a statistically reliable and valid measure of customer-bank attachment. We have shown that attachment is bi-directional. Customers become attached to service providers for a number of positive reasons including faith in organisational competence, mutually aligned and congruent values, and to experience relational value that is obtained from the interpersonal dynamics of their service encounters. We are committed to testing the applicability of the scale in other service organisation contexts.\n",
"role": "user"
},
{
"content": "Purpose: - This research paper seeks to investigate the different types of attachment that customers develop towards retail banks.\nDesign/methodology/approach: - The approach blends literature reviews, in addition to qualitative and quantitative methods. The primary research programme analyses data from seven focus groups, 39 one-on-one interviews and 1,058 individual consumer questionnaires.\nFindings: - Identifies three forms of positive attachment, based on three different foundations: the credibility of the organisation, compatibility between the values of the organisation and those of the consumer, and interpersonal or relational considerations. Presents a statistically valid and reliable scale that can be used to measure these three forms of attachment. Also identifies other forms of customer-bank attachment grounded on less positive associations, including inertia and alienation.\nOriginality/value: - The scale can be used to identify customer-organisational attachment profiles that transcend or complement customer satisfaction, and provide a basis for relationship longevity.\n",
"role": "assistant"
}
] |
The supply chain management of shopper marketing as viewed through a service ecosystem lens
|
[
"Knowledge management",
"Supply chain management",
"Agility",
"S-D logic",
"Service ecosystems",
"Shopper marketing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: A phenomenon within the consumer goods-retailing space, called \"shopper marketing\", has gained significant momentum in recent years and yet few academics, including supply chain management and marketing scholars, are paying much attention to it. Literally millions of marketing and retail-targeted trade dollars are shifting to shopper marketing strategies and initiatives and the implications for supply chain management, brand management and retail marketing strategy are significant. Having been described as a new discipline in the fast moving consumer goods space only recently, i.e. 2004-2007 (GMA/Deloitte, 2009), shopper marketing has fundamentally altered the way in which brand manufacturers, retailers, and all supporting organizations work together to develop strategic marketing and distribution programs. As a result, it has significantly altered communication along the entire path-to-purchase from out of store through to store merchandising, thus changing shopping experiences. Although shopping behavior (Underhill, 2009) and merchandising have been research foci for decades, only recently has scholastic attention turned toward shopper marketing specifically (Shankar et al., 2011). This \"new discipline\" has arrived with its own set of thorny challenges, some of which lie within the supply chain management domain. In fact, \"winning\" in this space is predominantly a supply chain opportunity as much as it is marketing.\nShopper marketing: Shopper marketing emerged most powerfully around 2004 when highlighted by the Coca-Cola Retailing Research Council of North America. Although numerous organizations from consultants to manufacturers debate the definition of shopper marketing, a good example is:\nS-D logic and service ecosystems: S-D logic brings core definitions to the issue and makes figural what is often left in the background. The concept of service ecosystems then adds a few critical concepts that reveal aspects of a shopper marketing network that make clear the knowledge management and supply chain imperatives for the future.\nInsights for shopper marketing and SCM: The S-D logic perspective of mutual service providing, resource integrating, density seeking, contextually specified, value co-creating actors that are endogenous to a value network suggests that all actors are a part of each other actor\nResearch directions: If we think of supply chain management challenges inherent within shopper marketing through the S-D logic and service ecosystem lenses, a number of research programs become evident:\nConclusion: Supply chain management, as it is traditionally conceptualized, includes processes that span \"organizations\" in support of \"goods\" transfer, storage and forecasting. Shopper marketing traditionally is conceptualized as a strategic relationship between \"brand manufacturers\" and \"retailers\" to extend \"goods\" brand equity built outside of the \"retailer\" all the way to the shelf in order to help the \"shopper\" find and purchase the \"goods\" she has a predisposition to buy quickly and easily. This enables her to solve the larger \"consumer\" problems she is trying to solve in her life, consistently, through her loyalty to specific brands and specific retailers. Supply chain managers are challenged with facilitating the transfer and storage of the correct resources at the right time, in the right place, in the right quantity in a new era of increased variance and complexity. This variance creates numerous within-organization resource integration challenges as well as inter-organizational ones. As Nonaka et al. (2008) point out, note that the actors in this ecosystem are human beings within units we refer to as \"organizations\". Management of knowledge flow among actors within and across organizations is critical. For example, the social networks of \"front line\" service employees of a distribution network, i.e. those executing shopper marketing initiatives within stores on behalf of brand manufacturers, must be better understood. These inter-personal social networks of front line employees extend deep into their own \"supplier\" organizations as well as beyond, can be quite complex, are largely about actors improving their resource density, and have a significant impact on in-store execution (Stolze, 2012).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine shopper marketing through service-dominant logic and service ecosystem lenses. In doing so, the authors reveal challenges and opportunities for supply chain management.\nDesign/methodology/approach: The work is conceptual, drawing on contemporary service-dominant logic thinking.\nFindings: Examination of shopper marketing reveals that it is currently stuck in goods-dominant logic and micro-level ways of thinking. By taking a macro service ecosystem view, all actors, including shoppers, are seen as resource integrators seeking resource density. The macro view highlights a significant amount of goods and information flow and variance now being added throughout shopper marketing systems.\nResearch limitations/implications: A guiding framework with appropriate terms defined offers new research directions and new ways practitioners can approach challenges in the industry. Research programs are suggested in the areas of facilitating resource density, examining the extent of ecosystems, measurement, mapping of resources, and creating shopper marketing innovations.\nPractical implications: This study provides an alternative way of looking at problems that arise in supply chain management planning and execution of shopper marketing initiatives.\nOriginality/value: Few scholastic articles address shopper marketing even within marketing and essentially none do so in supply chain management despite it having significantly disrupted supply chains since 2004. This article offers an overview of shopper marketing and helps supply chain managers identify quickly how they can add value and supply chain management researchers begin to address the challenges.\n",
"role": "assistant"
}
] |
The relationship between managerial entrenchment, earnings management and firm innovation
|
[
"Innovation",
"Earnings management",
"Corporate governance",
"Managerial entrenchment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Currently, earnings management is one of the central issues in accounting studies. Since the amount of profit is one main factor of decision making for investors, from a behavioral perspective these studies have their own significance. The results of these projects showed that low volatility and profit sustainability are two signs of quality. Hence, investors invest in those projects more confidently that their profit trend is more stable. When companies in unpropitious economic status are under an increasing pressure, managers ask the accounting department to improve the benefit and change the information content. Accounting despite all its flexibilities does not seem to be able to provide useful information for management in such circumstances (Hope and Hope, 1996). Earnings management usually occurs through manipulating the discretionary accruals or the real activities. Real activity-based earnings management includes the manipulation of real operation of a business unit in order to distinguish the reported profit of a current period (Mitra et al., 2013). The manipulation of discretionary accruals bears some costs. When the earnings management goes up, managers face the risk of investigation by auditors and lawmakers and are prone to litigation. The question raises here is that by the presence of agency problems, including information asymmetry and the ethical risk of investors and regulatory authorities, how we could be ensured of the quality of reported profit by the managers and in a broader sense the quality of the financial reporting?\n2. Theoretical principles, hypothesis development and literature: Prior to the development of capital markets, business firms were operated traditionally. These firms supplied their required capital through the properties of people and their entrepreneur relatives and the approach of founders was based on cooperation not investment. The factor of unlimited responsibility was a barrier to the partnership of individuals in investments. Joint-stock companies proposed from 1,855 upward. The advent of such companies is among the world's major economic evolutions and the separation of ownership from control is one of the consequences of the phenomenon ( Sajadi, 2009). The process of separating ownership from control takes place when stock ownership dispersion occurs due to the growth of a firm, the result of which is the decline of shareholders' power and the incremental enhancement of managers' authorities. The separation of ownership from control created the problem of manager accountability, as the agent of owners, and caused the shareholders not to influence the managerial section of the firm. This issue formed the basis of the agency theory (Sajadi, 2009). The agency relation based on the definition of Jensen and Meckling (1976) is a contract through which the employer or owner appoints an agent on their behalf and delegate the decision-making authority. Through the agency relation, it is assumed that each party is trying to maximize its own interests. According to the agency theory, the separation of ownership role from management leads to broker-agent demand, because it is likely that managers pursue their own interests even to the detriment of brokers (Mustapha and Che Ahmad, 2011). Corporate governance system is one of the available mechanisms to curtail the agency problem and information asymmetry between shareholders and managers in the capital markets (Shleifer and Vishny, 1986; Salehi, Tarighi and Safdari, 2018). This study evaluates the mutual effect of a set of corporate governance components and motivational factors on earnings management and firm innovation. Since the components and strategic corporate factors are analyzed in numerous studies, we are concerned about the relationship between these variables, in form of entrenchment indexes, and dependent variables of earnings management and innovative production of firms.\n3. Research method: In terms of objective, the present study is practical and in terms of data-analysis method, the study is cross-sectional correlational. Since the study is about the relationship between managerial entrenchment, earnings management and product innovation among companies listed on Tehran Stock Exchange, the desired method for hypothesis testing is a retrospective correlation, so the present study is a type of retrospective research. The information about independent, dependent and control variables was collected from the financial statements of the companies listed on TSE via reliable resources. The time range of the study was (2011-2016) as long as six years.\n4. Results: 4.1 Descriptive statistics\n5. Discussion: Consistent with agency theory, different corporate governance mechanisms in the form of managerial entrenchment led to reducing the manipulation of financial reporting. In this regard, our findings will warn investors and stakeholders that managerial entrenchment might be the best way in decreasing the earnings management and the agency problems in emerging economies, especially those markets struggling with financial sanctions like Iran.\n6. Conclusion: Due to the importance of the existence of the necessary mechanisms and the desirable quality of profit, most societies have made efforts to provide the appropriate corporate governance system. In recent decades, there have been different theories about the relationship between corporate governance mechanisms and profit management, and various empirical research studies have been carried out on these theories that led to the confirmation or rejection of the aforementioned theories. This research also explores the relationship between the quality of corporate governance in the form of the \"managerial entrenchment\" and earnings management among the firms listed on Tehran Stock Exchange. The results of the study indicate a negative and significant relationship between managerial entrenchment and accrual-based earnings management. This means that the entrenched managers are less likely to engage in manipulating discretionary accruals in the Iranian context. The second purpose of this study was to examine the link between managerial entrenchment and real earnings management using three different measures. In this paper, the measures of abnormal operating cash flow (CFO), abnormal production costs (PROD) and abnormal optional costs (DISCEXP) were used as a proxy for evaluating the real activities manipulation. The results witnessed that sanctuary managers are really interested in decreasing the level of activities related to real profit management in Iranian market. Hence, our findings are consistent with the studies of Pugh et al. (1992), Stein (1988), Gompers et al. (2003), Zhao and Chen (2008), Bebchuk et al. (2009) and Banko et al. (2013).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to evaluate the qualitative effect of corporate governance components, in the form of managerial entrenchment index, on earnings management and innovation.\nDesign/methodology/approach: In this study, the variable of managerial entrenchment, which includes the variables of management independence, dual role of management, management tenure, the board compensation and the board ownership percentage, was initially estimated through the exploratory factor analysis and its effect was evaluated on the dependent variables of the study using the test of multivariable regressions. Hence, a total of 103 listed companies on the Tehran Stock Exchange were selected and analyzed during 2011-2016. In this paper, the Jones model is used as the variable of accrued earnings management and for calculating the real earnings management, the models of abnormal operational cash flows, abnormal production costs and abnormal optional costs are employed. Moreover, the research and development cost to total costs ratio is used for calculating the innovation.\nFindings: The results indicate a negative and significant relationship between managerial entrenchment and accrual-based earnings management; moreover, the entrenched managers are less likely to engage in manipulating the real activities accruals in Iran context. Furthermore, the findings show that there is a positive and significant relationship between managerial entrenchment and firm innovation.\nOriginality/value: What really sets this paper apart from other studies is that this research will make aware investors and stakeholders of this fact that managerial entrenchment will be a good way to diminish the manipulation of financial reporting and improve the corporate situation in emerging markets, particularly those bazaars facing with economic sanctions such as Iran. Undeniably, the study results will complete the knowledge gap between the developed economies and the emerging markets.\n",
"role": "assistant"
}
] |
Reflection through web discussions: assessing nursing students' work-based learning
|
[
"Worldwide web",
"Communication technologies",
"Workplace learning",
"Action learning",
"Nursing",
"Finland"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: One of the questions in vocational training of today is how to keep the training from becoming too theoretical. Most approaches include shorter or longer work-based learning or work practice sessions in companies and other real-life work units; the complex nature of vocational skills and knowledge is best met by a comprehensive and many-sided approach to the assessment of work-based learning (Bjork, 2000; Docky, 2001; Flanagan et al., 2000). The assessment of nursing students in clinical practice is acknowledged as a long-standing and difficult problem. Vocational competence is difficult to define when applied to clinical nursing practice, and therefore difficult to assess.The internet technology offers multiple possibilities for learning (Parikh and Verma, 2002) and developing the quality of learning, reflection and cooperative knowledge building (Hakkinen and Arvaja, 1999; Midouser et al., 2000). This type of technology is especially promising when the learners are geographically distributed (such as work practice during formal training; distributed product development teams; multi-company or multinational networked project teams, etc.). However, the feasibility of internet technology for the assessment of learning has not yet been studied intensively. The present study investigates the evidence for the use of web discussion in the assessment of work-based learning in practical nurse education. The main interest is to analyse what kind of conclusions can be drawn from the students' learning and learning processes in general by following-up their self-reflection in a moderated web discussion.\nLiterature framework: Work-based learning develops more broadly transferable knowledge and skills (Stern and Rahn, 1998). It enhances the integration of theory and practice and makes authentic reflection possible (Lasonen, 2001). Burns (1994) argues that equal credit must be given to the assessment of both theoretical and experiential learning. This will only be possible if the process, as well as the outcomes of learning, is assessed.Difficulties to find consensus have made the assessment of clinical practice a mainly subjective process (Chambers, 1998). Watson et al. (2000) and Bjork (2000) have argued that the assessment of clinical competence remains problematic, superficial and incidental due to lack of rigour in the instruments and methods used for assessment. The process of learning can be measured by a judicious use of learning contracts, which incorporate a reflective element, allowing students to reflect upon and analyse particular clinical experiences, thus demonstrating their clinical competence.An important facet of continuous assessment is that of learner self-assessment, in which learners are required to judge their own performance. Such self-assessments could be used as efficient and effective learning tools, which allow the learner to identify his or her own strengths and weaknesses (Chambers, 1998).\nResearch methods and data: The sample in this study is from a development project in a Finnish vocational educational institution carried out by nursing teachers during the years 2002-2004. The authors conducted the research and analysed the data for research purposes. Eleven teachers tutored and assessed practical nurse students' work-based learning periods and took part in the action-learning project. The Finnish degree of Practical nurse includes 29 weeks of work-based learning in real hospital environments. Altogether 269 students from all the phases of their studies participated in the action research. The groups whose self-reflective web discussions were followed included two to 28 members, depending on the phase and vocational specialization of their studies. The age range of the participants was from 16 up to 43.Each student group took part in a web discussion between the study group and the tutoring teacher during all their work-based learning periods. The tutoring teacher visited the workplace twice during each period. The purpose of the latter visit was to negotiate the fulfillment of the student's individual targets based on her self-evaluation and the observations of the workplace tutor. The tutoring teacher took into account in her assessment the student's participation in the web discussions and her individual paper-based portfolio.The 13 web discussions from the first year of the study were tutored according to the principle of authentic dialogue, where the students were free to deal with their chosen topics, based on their experiences from the work-based learning periods (Aarnio, 1999). The authentic dialogue was found to be too open for practical assessment purposes and, therefore, the web discussions were slightly more structured during the second year.The authors analysed the web discussion data for research purposes with the NVivo 1.3 qualitative data analysis software. It was coded inductively, according to the principles of hierarchical coding (Richards and Richards, 1999). The purpose was to study how the processes of vocational growth, learning and reflection can be detected in the web discussion.The data consists of 26 separate web discussions, 26 qualitative, structured evaluation reports about each student group and the teachers' experiences written by the participating teachers. The students' web discussion data included altogether 8,770 separate messages. The discussion data of the smaller groups included altogether 13-120 messages and the discussion data of the larger groups between 568-804 messages.\nFindings and discussion: The results show that web discussion is a good media to reveal and demonstrate the process and outcomes of students' learning for assessment. Common discussion reached a more professional level and at the same time the differences between students became more visible. The quality and amount of students' messages reflecting upon their own learning, learning possibilities and vocational growth depended on the student groups' phase of study. The first-year students commented their own learning process and level of learning outcomes less than the more advanced students. The share of messages concerning students' own learning and vocational growth, counted as percentage of all the messages, was 33 per cent during the first year, 49 per cent during the second year and 59 per cent during the third year.\nExamples: reflection on one's own learning: Web discussion, students year 1\nExamples: reflection on the workplace practices: Web discussion, year 1\nConclusions: Structured web discussion is clearly a feasible method, especially over lengthy periods of time, for self-reflection and the documentation gathered can be used to analyse the learning process of individuals and teams, but is not alone adequate for assessing competence and work-based learning. The most important advantage is the possibility to follow the student groups' learning in real time. The fact that the students were aware of the impact of web discussion on their records certainly increased the students' motivation to participate in the common discussion and the exchange of knowledge. On the other hand, Cascio (2001) points out that students need to be encouraged to look upon the process as one that can potentially play a significant role in their professional development.The ability and activity to introduce new issues in the web discussion, which also increases other students' willingness to comment, is crucial for learning. It seems that the level of common and deep learning in the web discussion is reached only in the final phases of the vocational studies when common professional concepts facilitate common knowledge building. Thus the importance of web discussion in assessment must be lower in the first years of study and other methods, better suitable for documentation of individual processes, should be given more weight. Chambers (1998) argued that reflective practice is an important facet in the development of clinical competence and that competence may be assessed at least partially by assessing reflective skills. Thus web discussion serves best the assessment needs of the final phases of studies.The teachers participating in the action-learning project found web-based tutoring feasible and convenient. The web discussions dealt all the time with work and real practices. The teachers got documented and updated information about the students' learning in the workplaces far from the college. They recommended other teachers to use the web discussions as one assessment form and to make use of it also after the work-based learning period as an interesting, rich and authentic teaching material. Web discussion provides valid and valuable information about the discussants' learning and also about the contexts of it.For the students the web discussion was a new possibility for reflection on action (Schon, 1987), in which the common reflection also activates individual learning. The teacher's duty is to guide students into reflective thinking by forming questions. The technical feasibility to compare and analyse the self-reflective speech of the individuals and groups, for example the search facilities of the web forums, can largely affect the convenience of the assessment of each individual. It is crucial to analyse the contents and not only mechanically count the amount of individual students' messages. It seems to be a good practice that the persons who teach the student groups at a vocational institution also tutor their web discussion during work-based learning periods. Similarly, project managers should actively participate in and moderate the reflective web discussions of distributed projects.Although the ability to reflect is seen as the prerequisite for work-based learning, the vocational secondary level students do not self-evidently possess it. In our study, even after the third year's studies, not all the students have been able to develop the necessary reflection skills. Also Burrows (1995) indicates that the skills of reflection may be difficult to develop. Chambers (1998) argued that if reflection can be learned, then presumably it can (and should) be taught. Still, the practical nurse degree covers a pragmatic and practice-based occupation and the expectations from the students' competence should remain reasonable. The low share of comments about their own learning does not necessarily mean that the students have not been learning at all. More than that, students may lack consciousness of their learning when there is so much to be perceived and learned and they also have a totally new learning environment. Again, this situation can be very similar in entrepreneurial or other workplace learning environments (e.g. reflective practices after projects, in order to learn from them to the next project). There are typically numerous details in work situations that affect the outcome, and time seems to be always limited.Finally, the assessment of reflective practice through self-reflective discussions is one possible solution to the complex problem of the assessment of competence. The reflective practice fills the theory-practice gap. Reflection on action has been criticized for being narrow and distant from the real practice (Seibert, 1996). Chambers (1998) and Bjork (2000) point out that the ability to think and write does not necessarily indicate competent clinical practice. According to James and Clarke (1994), there is little or no evidence to suggest that such writing does indeed develop the reflective skills of students. Wong et al.'s (1997) study that utilized the models of reflective thinking of Mezirow (1991) and Boud et al. (1985) concludes that the written journal is a valid assessment of the learner's reflective skills. Based on our evidence, it seems that web discussion is a valid measure for the ability for reflective thinking and a useful method to develop reflection skills. The validity and reliability of the tools designed to actually measure the learning is, however, problematic and the truly objective assessment tool has yet to be developed (Chambers, 1998).\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to investigate evidence for the use of web discussion in the assessment of work-based learning in practical nurse education. The paper seeks to investigate what kind of conclusions can be drawn from the students' learning and learning processes, especially from their ability to reflect by analysing web discussions.\nDesign/methodology/approach: - The data were gathered in an action research project in a vocational secondary level institution. The data consists of 26 web discussions among student groups and their teachers' qualitative, evaluative reports, 26 altogether. The data were coded inductively with the NVivo 1.3 qualitative data analysis software. The analysis followed the principles of hierarchical coding. The purpose was to study how the processes of vocational growth, learning and reflection can be detected in the web discussion.\nFindings: - The paper finds that web discussion provides valuable information about the learning processes of students. It is a valid measure for the ability to reflective thinking and a useful method to develop reflection skills. Web discussion is feasible, but not an adequate assessment form of work-based learning all alone. It serves best the assessment needs of the final phases of studying.\nResearch limitations/implications: - The research was conducted in practical nurse education where communication and reflection are critical. More detailed research is needed from other vocational sectors.\nPractical implications: - The research findings suggest the use of web discussion as one valuable solution for the assessment of students' competence gained through work-based learning.\nOriginality/value: - The paper is based on original data.\n",
"role": "assistant"
}
] |
Scientometric analysis of research on "remotely piloted aircraft": A research agenda for the construction industry
|
[
"Innovation",
"Research agenda",
"Drones",
"Unmanned aerial vehicles",
"UAVs",
"Digitalization",
"Construction equipment",
"Information and communication technology (ICT) applications",
"UAS",
"Digital engineering"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Remotely piloted aircraft (RPA) are a type of unmanned aerial vehicle (UAV) (CASA, 2016). UAVs, popularly known as drones, are defined as \"any aerial vehicle that does not rely on an on-board human operator for flight, either autonomously or remotely operated\" (Rao et al., 2016, p. 84). Due to the technological advancements in miniaturization which have led to the increased availability of smaller, lighter and cheaper aircraft, the commercial use of RPA is rising (Padua et al., 2017). The construction sector has also embraced the use of RPAs (Gheisari and Esmaeili, 2019); in laying out sites, conducting remote observations and mapping, site inspections and safety monitoring, as well as inspection of the construction and operation phases of a wide range of civil infrastructure systems (Ham et al., 2016; Melo et al., 2017). Moreover, virtual simulation modeling, also an emerging field in construction management, benefits from data collected through RPAs (Han and Golparvar-Fard, 2017; Shakhatreh et al., 2018). In fact, RPAs are revolutionizing a wide range of construction activities through their advanced data collections capabilites. RPA-based methods surpass conventional methods on construction sites in terms of accuracy, efficiency and cost-effectiveness (Greenwood et al., 2019).\nBackground: The evolution of aircraft has advanced considerably since the launch of the first piloted aircraft, a hot air balloon, by the Montgolfier brothers in 1795. Over a century later in 1916, the Hewitt-Sperry automatic airplane became the first UAV (Zaloga, 2011), initially developed to deliver explosives in wartime (Clarke, 2014). UAVs have since been incorporated into a variety of commercial applications. According to the Civil Aviation Safety Authority of Australia (CASA, 2016), UAVs can be categorized into two types: RPA, which are used by government and private enterprise for practical purposes; and model aircraft, which are used recreationally. The distinction is a matter of whether there is a fee charged for use (Clarke, 2014). As RPA is the term applicable to the commercial application of UAVs on construction sites, it is the term used within this study. Nevertheless, other sources have offered other classifications of UAVs and interested readers are referred to Clarke (2014) and Shakhatreh et al. (2018) for details.\nResearch methodology: This study utilizes a \"mixed methods systematic review,\" as termed by Harden and Thomas (2010), as its approach. This approach applies and combines both quantitative and qualitative methods to study the literature according to a systematic procedure, with the aim of overcoming the potential bias of the subjective judgment inherent in mono-method manual systematic reviews (Harden and Thomas, 2010; Jin et al., 2018; Jin et al., 2019). As well as objectivity, the mixed-methods systematic review also provides enhanced \"depth and breadth\" (Heyvaert et al., 2016).\nRetrieved data results: Wave of research on RPAs\nThe research agenda: RPAs are a relatively new innovation and consequently have only recently been added to the arsenal of resources utilized by industry. Initial use was largely experimental, testing the range, reliability and quality of RPA capability. This piloting phase is reflected in the earliest research, beginning around 2012. But it was not until 2016 that the number of annual publications became significant, reaching double digits, signaling that RPA research had come into its own as an independent research discipline. Still, research remains in a fledgling state and limitations in the developing body of knowledge are to be expected.\nConclusion: RPAs are a relatively recent tool available to the construction industry and their long-term impact is set to be extensive. Research in the area is at an early stage, but growing rapidly. The subject of the inquiry undertaken was to examine what is currently being studied, the integrity of that research and importantly, what areas are being neglected. Limited, selective and subjective research into the extant breadth of knowledge on RPAs exists. This current study addresses this by applying a strong, comprehensive, mixed-methods quantitative scientometric approach to present the first systematic study on RPA literature.\n",
"role": "user"
},
{
"content": "Purpose: Remotely piloted aircraft (RPA) systems have emerged as an established tool within the construction industry. Concurrent with this trend has been the rise in research on RPA, establishing this as a new field of study within the construction management domain. What is needed now is an assessment of the current state of research in this emerging discipline - its strengths and weaknesses - by which future research on RPA in construction may be guided. The purpose of this paper is to address this need.\nDesign/methodology/approach: A total of 59 peer-reviewed journal articles covering RPAs within the construction domain were systematically reviewed using a mixed-methods approach, utilizing qualitative-scientometric analyses techniques.\nFindings: The results reveal a field of study in its fledgling stage, with a limited number of experts operating somewhat in isolation, from a limited number of institutions. Key publication outlets are identified, with the main focus of research being in the technical areas of remote sensing, photogrammetry and image processing.\nPractical implications: The study benefits researchers and industry practitioners alike. For researchers, the identified gaps reveal areas of high priority in future research. For construction companies, particularly small to medium-sized businesses, the study raises awareness of the latest developments and potential applicability of RPAs in the industry.\nOriginality/value: The study exposes what is missing from current research: a broader consideration of organizational adjustments needed to accommodate RPA usage, economic analyses and impediments to wider acceptance.\n",
"role": "assistant"
}
] |
International students' engagement in their university's social media: An exploratory study
|
[
"Higher education marketing",
"Social media",
"Customer engagement",
"Social identity",
"International students",
"Student engagement"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Higher education is an increasingly internationalised, competitive services sector and institutions now encounter unique challenges (Mazzarol and Soutar, 2012). One such challenge is managing student satisfaction, retention and loyalty. Higher education institutions invest significant resources in undertaking student engagement programs (Zepke, 2015), ensuring quality services are provided (Hennig-Thurau et al., 2001) and carrying out marketing activities to enhance brand equity (Mazzarol and Soutar, 2012). Many institutions use social media in such programs to facilitate the co-creation of learning experiences (Junco et al., 2013), improve student adjustment (Madge et al., 2009) and develop brand engagement (Belanger et al., 2014).\nSocial media and CE: Social media is \"a group of Internet-based applications that build on the ideological and technological foundations of Web 2.0 and that allow the creation and exchange of User Generated Content\" (Kaplan and Haenlein, 2010, p. 61). Tools include social networking sites, such as Facebook, Twitter and LinkedIn, blogs, customer review sites and online forums. Social media can be characterised as \"digital\" (high reach at low cost), \"pro-active\" (allowing users to take an active role in an organisation/brand process), \"visible\" (allowing users and organisations/brands to observe engagement/disengagement behaviours), \"real-time and memory\" (allowing the consumption and sharing of synchronous user experiences and archival information), \"ubiquitous\" (allowing interactions anytime and anywhere), and \"networks\" (allowing users to interact with others) (Hennig-Thurau et al., 2010, p. 13). These features enable organisations to have reciprocal interactions with stakeholders (Mollen and Wilson, 2010), empowering people to influence brand meanings and build community relationships (Schau et al., 2009).\nSocial media and social identity: While previous studies suggest that identification with a brand can drive CE in that brand's online community (Muniz and O'Guinn, 2001), recent studies have also found that online CE can enhance relationships with a brand (Hollebeek et al., 2014), perhaps because of the enduring nature of the CE process in which relationship qualities are strengthened by repeated engagement (Bowden, 2009). Thus, this research focused on the relationship implications of social media CE, as students' use of social media influences the formation of various social identities (Junco, 2014; Selwyn, 2009) and social capital (Ellison et al., 2007; Lin et al., 2012).\nMethodology: Since little is known about the phenomenon of interest, a case study of an Australian university was undertaken to explore international students' lived experiences with their university's social media. This approach allowed for the in-depth investigation of the contextual conditions that influence the contemporary phenomenon (Yin, 2009). Three semi-structured focus groups provided rich, holistic data from diverse groups of international students. By allowing participants to discuss these issues in a flexible way, the focus groups generated spontaneous responses that reflected genuine views, feelings, ideas and impressions (Sekaran and Bougie, 2013), while enabling the researchers to understand their norms, shared meanings and processes (Creswell, 2014).\nFindings: Engagement behaviours\nDiscussion: This research examined the nature of international students' engagement in a university-initiated social media outlet. Various aspects of the outlet's content and interactions influenced their engagement with the university and its student community. The students engaged with the site as part of their acculturation, although offline interactions still appeared to be of primary interest. Overall, the findings support the notion that computer-mediated communications can supplement and enhance place-based community engagement (Hampton and Wellman, 2003). Through multiple online and offline engagement experiences, students felt more emotionally committed and socially integrated to the university community than through their offline experiences alone.\nConclusions: Implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to understand the lived experiences of the international students using their university's social media, through a lens of customer engagement (CE) in the services marketing literature.\nDesign/methodology/approach: A case study was conducted in an Australian university. Three semi-structured focus groups with ten international students, along with a preliminary netnographic analysis of the university's social media account, provided a rich description of the phenomenon in the real-world context.\nFindings: The results suggest that these students are likely to engage in their university's social media as part of their acculturation and social identity construction strategy. Their engagement was cognitive and emotional, being influenced by the instrumental value of the social media page, engagement with campus rituals and artefacts, social identity and bonds with other students and perceptions of the page administrator. Furthermore, these students' engagement influenced their identification with the university and its student community, manifested in a sense of belonging and pride.\nResearch limitations/implications: The paper contributes to the higher education literature by offering relationship implications of social media CE. Limitations include small sample size and the single institutional context.\nPractical implications: The paper informs student communication practice, especially the design of university-initiated social media content and policy.\nOriginality/value: Universities and faculties today use social media to engage with students outside classrooms. However, little has been known about how international student sojourners view and respond to such initiatives. The paper addresses this gap by offering insight into how they engage with their university on social media and its relationship implications.\n",
"role": "assistant"
}
] |
Accounting for the male domination through legislative empowerment of upper-middle class women in the early nineteenth century Spain
|
[
"Gender",
"Spain",
"Household accounting",
"Langton-Dillón family",
"Patriarchy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Numerous studies have shown accounting to be an instrument of male domination of women in the public sphere, mainly in Anglo-Saxon professional settings (highlighted by Carrera et al., 2001; Cooper, 2001; Kirkham and Loft, 2001; Komori, 2008). Increasingly, researchers have also exposed the use of accounting practices to reproduce gender inequality in the private domestic arena (see Komori, 2012). Accounting in the private domestic arena has been shown not only to reflect social and cultural prescriptions but also to support the existing social order by reinforcing power relationships between men and women (Walker, 1998; Ezzamel, 2002a) as one of the \"less overt techniques\" of male domination (Walker, 2008, p. 591). In contrast, other studies have suggested that accounting has the potential to be a source of power for women when managing their households (Kirkham and Loft, 2001; Komori and Humphrey, 2000; Komori, 2012). This enabling power of everyday household accounting, until the present study, has not considered the implications when a woman has a legal entitlement to properties after marriage (Walker, 1998; Komori, 2007).\nFeminist social theory in the context: There is not one overarching theory capable of explaining gender divisions everywhere (Thane, 1992). Thus, different theoretical perspectives help to explain gender differences (Calas and Smircich, 2006; Haynes, 2008). This study gives particular emphasis to the perspective of feminist social theory which considers that women and men are subjected to socialization processes throughout their lives that result in different ways of viewing the world and in them behaving differently (Fischer et al., 1993). This perspective argues that gender differences are not only socially constructed and affect women's and men's perceptions and decisions, but also through the micro-processes that constitute social structures (Calas and Smircich, 2006) such as those occurring between individuals inside families.\nAccounting for gender: Accounting is not gender neutral but instead is both gendered and gendering (Carmona and Ezzamel, 2016). Studies by Komori (2008, 2012), Lehman (2012) and Carmona and Ezzamel (2016) provide useful insights into the current state of accounting research on gender and the need to deepen further our understanding of the interrelationship between accounting and gender. More specifically, until recently, the literature on accounting and gender has mainly focused on professional accounting occupations in predominantly Anglo-Saxon settings. Research has shown that particular occupations and practices may have gendered identities that can change over time and across different contexts (Kirkham and Loft, 1993; Roberts and Coutts, 1992). It has also been suggested that accounting can be a means of building a (non)gendered world (Parker, 2008). Several factors have been highlighted as reinforcing gender practices affecting the advancement of a woman's career in accounting, including: cultural values regarding the expected family roles of men and women (Anderson-Gough et al., 2005); stereotyping (Ciancanelli et al., 1990; Kim, 2004; Komori, 2008; Duff, 2011); dominant language and discourses (Cooper, 1992, 2001; Kim, 2008; Dambrin and Lambert, 2012); government regimes and religion (Carrera et al., 2001); and class, race and other specific circumstances of the women studied (Hammond and Oakes, 1992; Hammond, 1997; Gallhofer, 1998; Kim, 2004; Fearfull and Kamenou, 2006). Amidst this research, accounting remains under-researched as a gendering practice.\nMethodology: This work utilizes a qualitative research methodology based both on a historical interpretative perspective, sensitive to the context (Merino and Mayper, 1993), and on a longitudinal case study of a family to examine the role of accounting in the everyday lived experiences of women and men (Carmona and Ezzamel, 2016) and, hence, to \"make visible the informal and invisible processes of segregation\" (Calas and Smircich, 2006, p. 305). Recognizing that \"women\" is not a homogeneous category (Hammond and Preston, 1992), the Langton-Dillon family is viewed as an outstanding example of the Spanish upper-middle class at the time whose women had economic and property rights which were in direct contrast with those of women in the poorest classes. These unusual property rights required that the women had access to notarial and accounting information to verify and protect this patrimony (Ortego, 1999).\nThe Spanish context of the early nineteenth century: At the beginning of the nineteenth century, when the Ancien Regime still persisted, Spanish society was structured according to money and personal and kinship relationships. Under these circumstances, lineage or family interests prevailed over those of the individual, especially if this threatened the financial well-being of future generations with the subsequent loss of social status (Soria, 2007). In this patriarchal society, women were considered as physical, intellectual and morally inferior beings. Women were holders of their family's honor and, thus, their impoverishment could damage their family's reputation. This supposed weakness and vulnerability justified the deployment of a monitoring and protection system in which men's authority was unquestionable (Ortega, 1997). The contemporary law in force in Spain at the time, and more specifically in Castile where the Langtons lived, reflected the social structure of the time, thereby helping to reproduce it (Irigoyen, 2011). It reinforced the inferior status of women by preventing women from taking responsibility for their husband's debts.\nThe Langton-Dillon family: The first antecedents of the Langton-Dillon family in Spain can be traced to Lorenzo Careu, a successful trader from Ireland who had set up his business in Cadiz in the early years of the eighteenth century (Fannin, 2003). He had also sponsored the construction of a house in Ireland for poor widows who were admitted to live \"in seclusion and spiritual application for the service of God\" (Will of Lorenzo Careu drawn up in 1766). To protect the future of this charity, Lorenzo Careu bequeathed the usufruct of several estates in Cadiz. In addition, he left money to two hospitals for women, one located in Cadiz and the other in the nearby town of Sanlucar.\nAccounting for women's property rights: enforcing male domination: The female property rights emerging throughout MLC's first marriage became known through MLC's will given in 1798, when he specified his first wife's dowry as well as the value of her other personal patrimony. All these data, mainly obtained from accounting, allowed that he could calculated at that time the total amount that his first wife's only inheritor - his grandson Patricio Cruisse y Langton - should come into possession of. This amount should be delivered to Patricio when he got marriage or reached the full age, MLC having signed a promissory note before the notary for this concept the year before. This evidence shows the utilization of accounting to ensure that late woman's capital - initially preserved to protect her and calculated in detail - would returned to her own family group, flowing upwards or downwards.\nLangton-Dillon women's perception of accounting: Maria Dillon's will shows a woman's concern and self-perception about the importance of accounting for a married woman's future welfare in Castile. She repeatedly states in her will that she trusted that her husband's personal capital at the time of their marriage could be ascertained from MLC's accounting books, thereby allowing the calculation of the couple's common properties and the subsequent allocation of half this property to herself or her heirs: \"because of the full satisfaction that I have with respect to my husband's way to behave and his scrupulous conscience in money matters and for the same reason I am referring to the books themselves regarding his capital\" (Will of Maria Dillon, 1791). As a reputable traders' wife she trusted accounting as a technology to guarantee her own welfare and that of her lineage.\nAccounting as a gender-based technology in the household: Throughout his marriage, MLC recorded the money given to his wife in an account called household expenses (gastos de casa) which was kept as part of his mercantile accounting. The existence of this record reinforced the need for his wife to render accounts, thereby increasing her subjugation. This was also confirmed after the family moved to Bath when ML junior informed his father that all of the amounts withdrawn from the bank for his mother would be recorded in the same way.\nAccounting for defenceless women: From his arrival in Bath, ML junior kept the same accounting system as his father, charging the different family members for the cash used. After MLC died, his son continued to maintain the accounts of the different family members for the amounts of money that they had requested be withdrawn from the bank (see, for example, in Figure 4 the money withdrawn from MLC's bank account for Mari-a Dillon). This information was used by ML junior to manage the welfare of the women. In the case of Fanny Brun, the difficulties in recovering from her father money to cover her expenses, advanced to her by her grandmother, meant that ML junior resorted to accounting in order to safeguard Maria Dillon's welfare. To try and get Jose Brun (Fanny's father) to pay for his daughter's expenses, ML junior made reference both to Maria Dillon and her granddaughter's incomes and expenses, known by ML junior, and to Fanny's outstanding bills. After receiving a letter from Jose Brun, claiming that he would not pay for any of his daughter's bills until he knew the effective part corresponding to her in MLC's inheritance, ML junior wrote to him again, comparing Fanny's income with her own normal annual expenses:In order to refute that pretence [...] I have now sent him a copy of the (provisional) Repartimiento (sharing) [...] by which he will see that poor Fanny has no claim to more than four hundred and odd pounds upon the money in England [...] the interest of which in the 5% Navy stock after deducting the income tax is not more than PS20 a year whereas her expenses merely for dress, masters, & doctors, for no charge is made for maintenance, amounts to about PS70 a year, the balance of which my Mother disburses every year and he owes it to her and should pay it (letter to Agustin Butler explaining his last letter to Jose Brun, December 2, 1813).\nConclusions: This study has focused on the household accounting of a Catholic family of Irish ancestry, embedded in a very patriarchal context, who were representative of the Spanish upper-middle class in the early nineteenth century who lived for many years in the major trading city of Cadiz. At that time, a Spanish woman's dowry and other properties passed to her husband at the time of their marriage but could be recovered after the dissolution of a marriage, in contrast to Anglo-Saxon laws which withdrew a woman's property rights forever after she married. This difference in the treatment of married women has provided the opportunity to show the mutable character of patriarchal ideology and the accounting practices related to them (Walker, 2003).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to address both the neglect of non-Anglo-centric accounting gendered practices beyond the predominant professional setting and the controversial roles of women and accounting in power relationships inside the household. Analyzing a Spanish upper-middle class Catholic family in the early nineteenth century, the research focuses on the reciprocal interaction of accounting with practices and processes of daily life in a rigid patriarchal socio-cultural and juridical context.\nDesign/methodology/approach: This microhistory draws upon several archives, including in Spain the Archivo Historico Provincial de Cadiz. In England, the Bath Record Office has preserved documents and correspondence, both personal and business related, and the Worcester Record Office preserved notarial documents concerning the family. The large number of letters which have survived has facilitated an in-depth study of the people who were affected by accounting calculations.\nFindings: In a juridical context where women were conceived as merely the means for the circulation of property between two families, the evidence shows that accounting provided the proof of women's patrimony value and the means to facilitate their recovery in this cosification process. Although women had a little involvement in the household's accounting and management, they demonstrated confidence in accounting, fulfilling a stewardship function for the resources received. Also, evidence shows that by using accounting practices to shield supposedly defenseless women, this reinforced male domination over women and promoted the view that the role of women was as an ornament and in need of a good husband.\nOriginality/value: Contrasting with the Anglo-Saxon contemporary context, the Spanish law preserved a woman's property rights, guaranteeing recovery of properties owned by her before marriage should the marriage be legally annulled or be dissolved because one of the spouses' death. This required a detailed accounting of the wife's properties brought to her marriage, most especially regarding the dowry provided by her family.\n",
"role": "assistant"
}
] |
Google Scholar - friend or foe?
|
[
"Document delivery",
"Interlending",
"United Kingdom"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The library community has had a love/hate relationship with Google for a long time. Viewed by many of the general public as the ultimate one-stop shop for searching on any subject, end-users enthusiastically type one or two words into Google, hit return and become an instant expert. Despite library and information staff wearily pointing out that Google has weaknesses as well as strengths and might not, perhaps, be the best source of information in a particular field, it is still viewed as the source of wisdom by many people. Indeed, the very name of the company has been officially recognised as a synonym for online searching when it was added, in draft form, to the Oxford English Dictionary in June 2006 (Resource Shelf, n.d.). You can now officially \"google\" a search term.More specialist, in-depth information required by academic users, though, has continued to be catered for via various subject-specific resources that are carefully selected, managed and presented to the end-user by the library. That is, until November 2004 when Google launched a beta version of a new service - Google Scholar (GS). GS targets the academic community as a whole, from undergraduates to researchers. The aim is to cover a wide range of content providers and disciplines and to offer a single place where end-users can \"search across many disciplines and sources: peer-reviewed papers, theses, books, abstracts and articles, from academic publishers, professional societies, preprint repositories, universities and other scholarly organisations\" (Google, 2007a). Furthermore the service is free.Given that Google's (2007b) company mission statement is \"to organize the world's information and make it universally accessible and useful\", they could be seen as inherently challenging the role of the library. Google Book Search, also launched in 2004, has been a project to digitise books and make them available via another Google search facility. Although some libraries have partnered with Google in this venture, others in the community are more wary of the project and of where it might lead (MacColl, 2006; Banks, 2005). And GS could be seen as a similar challenge to the interlending and document supply services. Is GS directly challenging the role of the library as the hub where scholarly resources are discovered and delivered? Or is it offering a chance for libraries to work with a large partner who is a world leader in innovating and successfully introducing new technologies and thereby creating a better service for the end-user?So far, North American universities seem to have been on the front line of dealing with GS. But UK libraries are increasingly noticing that end-users with research interests, from students to the higher levels of researchers, are using GS. The time seems to be ripe to look at the service from a document delivery perspective. In the first part of this article, I intend to review and summarise the service and the main points of the current debate both in practical terms and also look at the bigger picture. In the second part of the article, to be published in a later edition of this journal, I intend to explore the ways that UK libraries are dealing with GS and to give a snapshot of what library staff see as the pros and cons of the service, both in practical and conceptual terms.In practical and conceptual terms, then, how does GS measure up on the discovery and delivery of scholarly materials? There are three main areas to look at:1. searching;2. access to the most appropriate copy; and3. integration with existing library services and technologies.\nSearching: The GS search has a familiar interface for end-users as it looks and feels like using the Google basic search. It is easy to navigate, simple to use and is fast at returning results. There is also an advanced search option very much in the Google style, so familiar to most people. The concept of offering one interface where end-users can carry out a federated search is appealing to end-users and again simple to use and understand. There is a citation-tracking tool that shows, for articles, where else the work has been cited, a nice piece of functionality in principle for people who are carrying out research. As for result ranking, GS claim: \"The most relevant results will always appear on the first page\" (Google, 2007c). So far so good - a simple, intuitive interface, one place to enter search terms, a basic and advanced search that already looks and feels familiar to the end-user; plus some attractive additional functionality.However, there are drawbacks. The main issue is exactly what GS is searching. Google is very coy about actually naming names on any of the information pages. In addition, although GS does search some respectable academic sources (e.g. Ingenta, CrossRef, OCLC WorldCat), other results are gathered by crawling the open web. This means that not only might your end-users not be accessing the best sources for their subject, but also they might simultaneously be exposed to hits returned from very dubious sources on the open web. The searching itself is not as comprehensive as most searches offered through academic sources by library catalogues, journal publishers and aggregators. The presentation of results has also been the subject of much criticism, for despite the claims that the most relevant results appear on the first page, there seems to be many instances of older, less relevant material being ranked above more recent publications due to the algorithm used by GS. Of course, the danger is that these weaknesses add up to dubious hits. And those dubious hits will appear legitimate when shown within an interface branded as scholarly. Even for end-users who are aware of the limitations of the main Google search, this could be very misleading.\nAccessing the most appropriate copy: Moving on to the holy grail of document delivery - the most appropriate copy - GS has tackled the issue of including the holdings of the home library physical collection and giving links to online full-text articles that the home library subscribes to on behalf of end-users. This means, in theory, that a library can encourage end-users to search via GS and still make sure that the most appropriate copy (an on-the-shelf hard copy, a link to a full-text article via a library-level subscription) is easily available. There are, though, some practical drawbacks. Libraries wanting to point users at their home library stock must have their holdings loaded in OCLC WorldCat. The end-user then sees a link to WorldCat next to all GS book results. Clicking on the link takes them to WorldCat where they will see a list of libraries local to them that hold the book (Google, 2007d). This has a number of steps involved and is not as immediate as actually seeing the home library holdings listed within the results set, a functionality that is available in most portals on the market. Likewise, access to full-text, subscription articles also involves a number of steps (Google, 2007e). It also requires the library to have a link resolver in place and seems to work best for on-campus end-users, although off-campus end-users are usually significant users of online resources in many academic environments. In this category, GS has certainly made an effort to respond to the requirements of the library, but it is still very much a work in progress with room for improvement.Interlending, though, is not considered at all. Most libraries would want to make the decision about how an item was sourced and delivered to the end-user if it was not available as either an on-the-shelf hard copy or accessible electronically as full text. The most common way to handle this situation is to channel end-users through the interlending process by prompting them to put in a request for the item. This is now often via an online request form or a message that is linked to items outside the library resources and informs end-users how to place a request. GS has no internal provision for requesting and does not offer an option to link into the home library's ILL system. Instead, end-users may be prompted to pay directly for access to a loanable item or a copy. This could be in direct contradiction to a subsidised or free-at-the-point-of-delivery policy offered by the home library through the ILL network. This is a serious flaw in GS and offers the end-user an inferior service for this scenario.\nIntegration with existing library services and technologies: This final category has partly been addressed above. GS does not yet give a smooth and fully integrated view of internal and external resources. For libraries, GS might present an even bigger stumbling block in the look, feel and branding it brings to end-users. The implications of this seem to be the elephant in the room in many critiques of the service, but are an important part of the debate. If end-users are working with GS, they see predominantly Google branding. They search in a Google interface, they handle results in a Google interface, they click on Google-style text links rather than their own branded button to reach full-text articles (Google, 2007f), even when those articles are available because the library has paid for a subscription to the content provider. It could easily start to look like the library services are an add-on to the GS service rather than the other way around.In addition, the shortcomings of GS can be all too easily overlooked by end-users because the interface is familiar, easy to use and delivers fast results that seem relevant. There is a challenge here for libraries that has already been identified in the USA (Giustini and Colenbrander, 2006). GS may have shortcomings and library-designed search interfaces may deliver better quality, more relevant and more appropriate results. But end-users are often not prepared to learn to use comparatively complex, unfriendly and difficult interfaces for searching.In essence, on a practical level, GS has some work to do, but has the potential to be a useful resource. However, the implications of the presentation and branding of the search environment raise some important concerns. On a conceptual level, the question is a little more difficult. GS raises some important questions for library-based document delivery services about the ongoing role of the library as the hub at the centre of this service. How libraries might react to the service seems to fall into two broad camps. Either embrace Google as a partner and work with the search giant to build GS into a service that suits library needs, or provide and promote in-house services to end-users and not endorse GS. Both views assume that the library will continue to have a pivotal role in the discovery and delivery of scholarly material and each argues that their chosen path is the most constructive way to develop their services into the future.In part two of this article, I will be looking at how both camps are fairing in practical terms and in their views on the implications of Google's interest in scholarly publications. If any readers would like to contribute their own experiences and opinions, I would be delighted to include them. A list of resources that go into the different areas of the debate in much more detail than has been possible here is included (over) to be used as a primer for those who would like to know more.\nResources: Information from Google\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to review and summarise the service offered by Google Scholar (GS).\nDesign/methodology/approach: - This is the first of two articles on GS, the second of which will explore the ways that UK libraries are dealing with Google. Both deal with the subject with particular reference to document supply.\nFindings: - The article finds that on a practical level, GS has some work to do, but has the potential to be a useful resource. However, the implications of the presentation and branding of the search environment raise some important concerns. On a conceptual level, the question is a little more difficult. GS raises some important questions for library-based document delivery services about the ongoing role of the library as the hub at the centre of this service.\nOriginality/value: - The article provides a useful overview of today's hot topic which includes a list of resources for further reading and help.\n",
"role": "assistant"
}
] |
Public procurement for innovation and civil preparedness: a policy-practice gap
|
[
"Innovation",
"Policy implementation",
"Preparedness",
"Public procurement",
"Security of supply"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Public authorities fulfil their mandate in part by procuring goods, services and works for the public good. Through the public procurement process, designed at achieving certain outcomes, public authorities can also bring about financial and monetary opportunities for industry and suppliers. In addition, procurement is not only a mechanism for acquiring products, but also a process argued to have \"outcomes and impacts that are policy-related ends in themselves\" (Atkinson and Sapat, 2012, p. 363). Public procurement has also been argued to be a tool for achieving or promoting policy goals such as innovation or industrial development (Borras and Edquist, 2013; Snider and Rendon, 2008).\n2. Conceptual framework: As stated already, the focus of this paper is to explore how the policy goals of innovation and preparedness trickle down to procurement practice - in other words, the paper is focussed on policy implementation. Public policy is defined by Peters (2013, p. 4) as \"the sum of government activities, whether pursued directly or through agents, as those activities have an influence on the lives of citizens\". Governments can use different types of instrument to implement policy. The term public policy instrument is defined by Vedung (1998, p. 21) as \"the set of techniques by which governmental authorities wield their power in attempting to ensure support and effect or prevent social change\". Vedung (1998) uses a threefold arrangement to categorise policy instruments, namely, regulations (sticks), economic means (carrots) and information (sermons). In this typology, regulations refer to coercion, economic means to the use of remuneration or deprivation of material resources, and sermons to intellectual and moral appeals (Vedung, 1998, p. 30).\n3. Methodology: 3.1 Data collection\n4. The policy-practice gap: 4.1 General overview of policy goals and instruments\n5. Discussion and conclusions: 5.1 Discussion\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to give a deeper understanding of public procurement for innovation and civil preparedness. The paper focuses on how the policy goals of innovation and preparedness are implemented in procurement practice, and explores the issues that affect innovation- and preparedness-oriented procurement.\nDesign/methodology/approach: The study is exploratory and based on empirical data from semi-structured interviews with 92 informants in the healthcare, energy and water services sectors in Finland. Interview data were analysed using a coding template, and further categorised to answer the research questions.\nFindings: The study brings to light how procurers perceive the different but overlapping policy goals of innovation and preparedness, and whether and how the policy goals are turned into outcomes in the procurement process.\nResearch limitations/implications: The data were collected from case sectors in Finland. However, considering the adoption of European Union legislation, some of the findings are generalisable in other sectors, organisations and countries.\nPractical implications: The paper highlights aspects that can aid procurers, suppliers and policy-makers in understanding how public procurement can be an instrument linking innovation and civil preparedness.\nSocial implications: It cannot be overemphasised that populations are exposed to more and, arguably, greater hazards. Public authorities can use innovative solutions and processes in meeting multiple objectives such as everyday societal needs as well as civil preparedness and planning.\nOriginality/value: Such research, purposely sampling case sectors providing critical services, has not been undertaken before in Finland. This work provides valuable information for organisations involved in public procurement, innovation and civil preparedness.\n",
"role": "assistant"
}
] |
An overview of six sigma applications in healthcare industry
|
[
"Six sigma",
"Health services"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Health care today is a vast web of complexity and contradiction. It offers astounding advances in technology and treatment, but is often overburdened by inefficiencies, errors, resource constraints and other issues that threaten the accessibility and safety of patient care. In 1998, the Institute of Medicine released an assessment stating that 98,000 people die each year as a result of medical errors, highlighting the necessity for quality improvements (Lazarus and Neely, 2003). An estimated PS400 million is being paid in clinical negligence claims and adverse incidents resulted in approximately PS2 billion per annum (Department of Health, 2001; Milligan and Robinson, 2003).Over the past decade, the need and desire for healthcare institutions to operate more efficiently has been driven largely by financial concerns. With decreasing reimbursements and the rising cost of labour and supplies, healthcare organizations have been driven to look within to find savings. Generally, the accepted rule has been that there are two primary areas to find savings, either through reductions in labour costs or by reducing the high cost of supplies. With the continuing shortage of qualified healthcare professionals, registered nurses, registered radiologic technologists, etc., the opportunities to reduce expenses solely by reducing labour costs can play an important role in causing bottlenecks in healthcare services.Today's health care organizations are complex dynamic systems. They are more focusing to improve quality of care and meet stringent guidelines. Thus, re-examining the method of evaluating the service performance has become necessary. All quality improvement activities should focus on improvements in:* clinical outcomes;* satisfaction; and* efficiency.These are determined by system characteristics.Improving levels of patient satisfaction is very critical to a healthcare organization's long-term success. To improve patient satisfaction, healthcare providers must focus on quality improvement strategies. That is, healthcare professionals must demonstrate attributes consistent with organizational culture. Six attributes for a quality healthcare system are identified as follows:1. Safe.2. Effective.3. Patient-centred.4. Timely.5. Efficient.6. Equitable.Monitoring these attributes is a crucial element of organization's philosophy and should be part of the quality improvement initiative.Much effort has in the past and in some places still is directed at imposing quality rather than concentrating on optimizing the healthcare system. Sigma methodology is an optimization tool that focuses on developing and delivering near-perfect services. It is one of the most powerful performance improvement methodologies that are changing the face of modern healthcare delivery. The DNA of the six sigma methodology can be felt while reducing variation that leads to safer, quicker and more co-ordinated care, less mortality and morbidity, better response to needs and better value from resources.Up-to-date, six sigma projects in healthcare industry have focused on direct care delivery, administrative support and financial administration (Antony et al., 2006). Six sigma projects can be executed in the following healthcare processes:* increasing capacity in X-ray rooms;* reducing avoidable emergency admissions;* improving day case performance;* improving accuracy of clinical coding;* improving patient satisfaction at Emergency Room (ER);* reducing turn around time in preparing medical reports;* reducing bottle necks in emergency departments;* reducing cycle time in various inpatient and outpatient diagnostic areas;* reducing number of medical errors and hence enhancing patient safety;* reducing patient falls;* reducing errors from high-risk medication;* reducing medication ordering and administration errors;* improving active management of personnel costs;* increasing productivity of healthcare personnel;* increasing accuracy of laboratory results;* increasing accuracy of billing processes and thereby reducing the number of billing errors;* improving bed availability across various departments in hospitals;* reducing number of post-operative wound infections and related wound problems;* improving MRI exam scheduling;* reducing lost MRI films;* improving turn-around time for pharmacy orders;* improving nurse or pharmacy technician recruitment;* improving operation room throughput;* increasing surgical capacity;* reducing length of stay in ER;* reducing ER diversions;* improving revenue cycle;* reducing inventory levels;* improving patient registration accuracy; and* improving employee retention.\nTheoretical background: High performance is essential in every healthcare process. Sigma is a Greek letter of the alphabet used to describe variability, or in mathematical terms, standard deviation. Six sigma offers a way of measuring the performance capability of existing systems or processes. It is a statistical unit of measure that reflects the likelihood that an error will occur. Six-sigma relies on rigorous statistical methods, and implements control mechanisms, in order to tie together quality, cost, process, people, and accountability. It begins with an understanding of customer requirements, and values (Riebling, 2005). The six-sigma goal is to reduce both variance and control processes in order to assure compliance with the critical specifications.The higher the sigma level, the higher the performance of the healthcare system. For example, a 3s process has a defect rate of 6.7 percent whereas a 6s process has 3.4 defects per million opportunities. An Institute of Medicine (IOM) report states that deaths attributable to anaesthesia have fallen to 5.4 per million, which approaches 5s and added that medical errors are responsible for the deaths of 44,000 to 98,000 hospital patients every year (Bion and Heffner, 2004). In addition, reducing medical errors can also result in significant financial savings, reporting that total national costs of preventable medical errors resulting in injury account for between $17 billion and $29 billion per year. The study noted that preventable adverse drug reactions can increase average hospitalization costs by $4,700 per admission (Bion and Heffner, 2004). Also, the average of doctor prescription writing has been found to be between three and four-sigma (Rudisill and Druley, 2004). These examples can be improved by means of applying the six-sigma philosophy.Organizations need to improve what they are currently doing as well as changing operations to what they should be doing, and after measuring the outcomes they should search for improvements on other new activities. Every improvement requires a change, either in small scale or in large scale, and every change requires an act of creation (Sommers, 1998).Patient satisfaction, physician satisfaction, reduced overtime, reduced patient wait times, increased revenues and an enhanced quality of life for healthcare personnel are some of the outcomes of moving to the higher sigma level. The goal is to move from the current state to a future, more productive state.By means of adopting six-sigma philosophy, the healthcare organization can achieve a cultural change. This will lead in realizing sustainable bottom-line results in the hospital. Having committed and being supportive, the top management must select quality strategies that tie projects to organization imperatives. Thus, it can identify meaningful performance metrics (see Figure 1).From emergency room to boardroom, six-sigma can reduce variability and waste by translating to fewer errors, better processes, improved patient care, greater patient satisfaction rates, and happier, more productive employees. To achieve these goals, the DMAIC must be implemented.The DMAIC is a five-step improvement cycle with the aim to continuously reduce errors: Define the project by identifying problems, clarifying scope, defining goals. Measure the current performance against requirements, gather and compare data, refine problems/goals. Analyse by developing hypotheses, identifying sources, gaps and root causes, analysing best practices. Improve by conducting experiments to eliminate root cause, testing solutions, measuring results, standardizing solutions, implementing new processes by designing creative solutions to fix and prevent problems. Control the performance of the process by institutionalizing improvements and putting a mechanism for ongoing monitoring in place.Healthcare organizations implementing six-sigma using the DMAIC methodology, should later add the DFSS (design for six sigma) methodologies when the organizational culture and experience level permits. However it is always advised to look at the improvement aspects of existing processes using DMAIC methodology and then apply Design for six sigma methodology either to re-design existing processes for further improvement or to design and develop a new service from scratch.At its core, six-sigma revolves around the following key concepts:* Critical to quality: Attributes most important to the patient.* Defect: Failing to deliver what the patient wants. In terms of impact to the patient, a defect in the delivery of healthcare can range from relatively minor to significant. In a worst-case scenario, the defect can be fatal, as when a medication error results in the patient's death.* Process capability: What the healthcare process can deliver.* Variation: What the patient sees and feels.* Stable operations: Ensuring consistent, predictable processes to improve what the patient sees and feels.* Design for six sigma: Designing to meet patient's needs and process capability.* Lean six sigma: integration of Lean Thinking (that means speed and better flow of the processes by eliminating waste) and Statistical Thinking (that means understanding data, process and variation in processes).\nSome key challenges and barriers in the implementation of six sigma in healthcare: The first and foremost challenge is the initial investment in six-sigma Belt System training. The absence or difficulty to obtain the baseline data on process performance is another major challenge while applying six-sigma in health care sector. There will be lots of data available in the health care sector, however, most of the time these data are not readily available for its analysis. For health care industry, it is often a struggle to identify processes, which can be measured in terms of defects or errors per million opportunities (Lanham and Maxson-Cooper, 2003). Another barrier to six-sigma deployment in health care industry is the psychology of the workforce. Last but not the least, it is important to present recommendations using the business language rather than the statistical language.\nSample applications of six sigma methodology in healthcare industry: In this paper, the following five case studies of six sigma applications are suggested:1. Decreasing unnecessary laboratory tests.2. Improving Magnetic Resonance (MR) image quality.3. Decreasing waiting times before surgery.4. Reducing catheter infections.5. Decreasing excess length of stay in hospitals.Case 1: Decreasing unnecessary laboratory tests\nConclusions: Healthcare industry is still in the early stages of evolution with regard to six-sigma. Therefore, healthcare personnel should seek guidance for training and implementation by the support of the top management. Successful execution of simple projects in hospitals can enable practitioners to tackle tougher initiatives in the future and create clinical change on a broad scale.Integrating the six-sigma culture into entire organizations by the commitment and involvement of top management can multiply the positive effects and make a significant impact at all levels. High level of internal communication is also necessary to facilitate the implementation of six sigma. The established Quality Improvement Team should collaborate with outside quality facilitators who can train them in deploying implementation and to provide necessary training for all project participants in system deployment, project management, and utilization of quality tools.Appropriately implemented, six sigma clearly produces benefits in terms of better operational efficiency, cost-effectiveness and higher process quality. In addition, it also has an impact in clinical areas such as infection control and medication delivery.Impeding factors to the implementation of six sigma in healthcare organizations are as follows:* lack of financial resources;* lack of human resources;* lack of time;* lack of leadership;* poor training;* poor project selection; and* internal resistance.The authors believe that six sigma as a business strategy allows health care sector to deliver a truly high-class service to patients. Think of the true impact that six sigma could have if we focus on the core issues of health care and improving the quality of lives of patients. In authors' opinion, the application of six sigma in health care industry will continue to grow, especially here in Europe over the next five years or so.\n",
"role": "user"
},
{
"content": "Purpose: - Delays, measurement and medical errors and variability often undermine the delivery of safe, effective patient care. However, it is possible to minimize them by applying six-sigma. This methodology aims to focus on the root causes of healthcare problems, analyses them by flowcharts and fishbone diagrams and produces near-perfect healthcare services.\nDesign/methodology/approach: - Five case studies in healthcare are designed to show the performance improvement accomplished by six sigma. The DMAIC (Define-Measure- Analyse-Improve-Control), i.e. a road-map for problem solving and service/process improvement, is implemented.\nFindings: - The findings in this paper show that the healthcare organization has a greater ability to address challenges across the system. Resource utilization has been maximized. Fewer redundancies, waste and rework have been observed. Bottle-necks related to scheduling have diminished. Working conditions have improved for healthcare personnel. Increased patient and physician satisfaction as well as cost savings have been achieved. These will enable the healthcare organization to increase its market share in the long run.\nOriginality/value: - The application of six sigma in healthcare services is relatively new topic and very little research has been performed in this area. The paper will be extremely valuable to researchers and practitioners who are currently engaged in six sigma research.\n",
"role": "assistant"
}
] |
Without borders: youth debaters reimagining the nature and purpose of public dialogue
|
[
"Critical literacy",
"Literacy and identity",
"English language arts"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The debate tournament was beginning. Pairings for the first round had been released and each team eagerly scanned the list: Are we aff or neg? Who are we up against? What room are we in? Students sought final words of encouragement from their coaches and then took off - don't want to keep the judge waiting. It was April, so every debater could recite the resolution by heart; after all, it was the same one that all middle and high school policy debaters had been working with all year long. Resolved: The United States federal government should substantially reduce its restrictions on legal immigration to the USA.\nYouth epistemologies disrupting debate: This article explores how the choices that students of the Debate Liberation League (DLL) made - the integration of spoken word performance, the fluid use of both Spanish and English and the critique of the norms of policy debate itself - represented a conscious effort to not only comment upon US immigration policy but also destabilize the boundaries of a normative literacy activity to demand recognition for more expansive epistemological resources. This exploration was structured around the research question: How do DLL students leverage their linguistic and cultural repertoires of practice (Gutierrez and Rogoff, 2003) to offer a counter-narrative to normative debate literacies? Through the course of this inquiry, our intergenerational, collaborative team of authors found that DLL students advocated for the insertion and amplification of their identities, family histories and linguistic resources into debates about immigration policy to construct a foundation for inclusive deliberative dialogue and to imagine a more equitable and justice-oriented academic and civic life. They resisted the assumption that they needed to minimize or ignore these cultural constructs to be considered ready or able to participate in a policy debate.\nNeoliberal narrative of debate: Debate is an academic literacy activity that has been heralded for its potential to expand student vocabulary, increase reading comprehension skills and foster critical thinking through its focus on persuasion, questioning and perspective-taking (Snow et al., 2009; Zwiers and Crawford, 2011). Because it often involves deliberation about controversial social issues and current events, debate is also considered a best practice of civic education (Gould et al., 2011; Hess and McAvoy, 2014). While it can take myriad forms, some of the principles that are often considered foundational to the practice include two sides (pro and con), competition before a judge who decides upon a \"winner,\" and the preference for credentialed sources over subjective or embodied knowledge (Bellon, 2000).\nCritical race counter-narrative of debate: In his memoir, White Like Me, writer and activist Wise (2005/2011) describes his experience as a policy debater in these terms:\nBorderlands and critical literacy: Being a principal involves tough compromises. I know that how schools define literacy is too narrow, but I also want my students to succeed in this pipeline. There is no ideological purity - it's more of a both/and.\nMethodology: This study is part of a broader research project investigating the impact of debate participation on middle school students' academic, socioemotional and civic development. The project emerged from a social design research partnership with the Middle School Quality Initiative (MSQI), a program of the New York City Department of Education dedicated to improving literacy outcomes and college trajectories of middle grade youth. The use of social design research reflects our commitment to collaborative inquiry and design, an explicit focus on equity and the co-construction of knowledge about shared problems of practice (Gutierrez and Jurow, 2016). We were specifically attuned in this study to the use of shifting language modalities and the enactment of identity that occurred in the context of literacy practices.\nResearcher positionality: Nicole's identity, both in terms of her specific relationship to the study participants and her broader socioeconomic positioning in US society are important elements to explain to ensure transparency in the discussion of analysis and findings. Nicole was formerly an English teacher and colleague of Southside Principal, Alex Brooks, at a different New York City high school, where the two of them coached the policy debate team and learned about critical performance debate. When Alex began working with MSQI, he helped broker the relationships to foster the research partnership from which this study emerged. And when he became the Principal of Southside, he invited Nicole to advise his new debate team. Nicole's visits with the team involved both assistance with the development of the team's debate performance and observations for the purpose of data collection.\nResearch context: Southside Middle School is a nonselective public middle school of 320 students in New York City that since 2014 has partnered with a local human rights and community advocacy organization to serve as the educational hub of its community. While it had been facing decreasing enrollment and struggles to make sufficient progress on the district's accountability benchmarks, recent years have seen increases in enrollment and achievement as Alex Brooks sought to make debate the focus of the school's turnaround efforts. An outlier in a rapidly gentrifying neighborhood, Southside's student body is 75% Latinx and 25% black. Eighteen per cent of students are designated as current English Language Learners (ELLs) (with another 30% considered former ELLs), and the school welcomes an average of one student per month who recently arrived to the city from Central American countries.\nData collection and analysis: Data for this study was collected during the 2018-2019 school year. Observational data included field notes from 15 debate practices (2 h each), as well as audio recordings of selected discussions and practice rounds (5 h). Interviews were conducted with the three adult mentors (30-60 min each) and two focus groups were conducted with the students (60 min each). Artifacts including student reflections/notes, debate case materials and poetry were gathered for analysis and Nicole wrote five analytic memos in consultation with the group throughout the school year.\nFindings: To address our research question regarding the ways in which DLL students leveraged their linguistic and cultural identities to challenge normative debate practices and reimagine the activity through their own epistemologies, we organized our findings into two sections that echo the movement of our theoretical framework from neoliberal to borderland conceptualizations of debate. We first explored how students came to identify and wrestle with debate norms that excluded their experiences and perspectives as they read and analyzed Anzaldua's theory during their debate practices. We then documented the moves they made in the development of their critical performance debate case to offer a counter-narrative that deconstructed those norms and rebuilt a more expansive literacy practice.\nDiscussion and implications for practice: Nuestra vernacula se infiltraen el espacio de debate y resiste el silenciamiento que ocurre dentro del status quo. Ya no permitiremos que nuestro discurso sea excluido dentro de este espacio!\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to analyze how a group of middle-school debaters integrated their identities and epistemologies into the traditional literacy practice of debate to advocate for more expansive and inclusive forms of academic and civic discussion. The adult and youth co-researchers of the Debate Liberation League (DLL) detail their creation of a critical debate praxis through the use of spoken word and translanguaging and illustrate how they sought to redesign a foundational activity of English Language Arts on their own terms.\nDesign/methodology/approach: Drawing upon critical race and borderlands theories, the authors use critical ethnographic and participatory action research methods to explore how the DLL deconstructed the boundaries of what counts as public dialogue and offered an alternative model of what intergenerational and multi-voiced democratic discourse could look like in English Language Arts (ELA) classrooms and beyond.\nFindings: The findings demonstrate how DLL students broke down normative binaries of affirmative/negative and objective/subjective in their debate performances and introduced testimonios as evidence for civic claims to make space for their voices and reimagine deliberation.\nOriginality/value: This study foregrounds dialogic data generation through a collaborative, intergenerational research approach. It highlights the constructed nature of literacy \"rules,\" demonstrates youth expertise in reimagining ELA, and offers a pathway toward a more compassionate public sphere.\n",
"role": "assistant"
}
] |
The Ombudsman Act 1980: redressing administrative grievances in Bangladesh
|
[
"Bangladesh",
"UK",
"Ombudsman",
"Government resistance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Since the 1960s, the concept of an ombudsman has developed rapidly in both the academic world and in more practical fields. Since then, this institution has spread all over the world, particularly in developed and developing countries. For these reasons, some scholars have described the institution of the ombudsman as \"being in fashion\" (Kastari, 1968, p. 58; Abedin, 1986). During the 1990s, the necessity of the Office of the Ombudsman became an acceptable institution in various countries of the world. Sir John Robertson (former Director of the International Ombudsman Institute) rightly remarks, \"We are in the midst of the most important and exciting decade of the twentieth century for the development of the ombudsman as an institution\" (Robertson, 1995). Now-a-days, the institution of the ombudsman is popular because of the timely resolution of grievances in various sectors, \"in particular government sectors and in the private sector [...] [e.g.] in corporations, industries, colleges, and universities\" (Reif, 1999, p. 13). The ombudsman model has also been:\nOmbudsman in Bangladesh - a historical note: Regrettably, there is no Office of the Ombudsman in Bangladesh, although the provision of ombudsman has been incorporated in the Constitution of Bangladesh in 1972, as per Article 77[1]. An attempt was made to effect the constitutional provision of Bangladesh, and the Ombudsman Act, 1980 (1980) for the adoption of the Office of the Ombudsman was passed during the regime of President Ziaur Rahman. However, after the assassination of President Ziaur Rahman, the succeeding governments, particularly President Ershad's nine years of dictatorship rule, did not establish an Office of the Ombudsman. After a long struggle against the Ershad regime, President Ziaur Rahman's party Bangladesh Nationalist Party (BNP) again came to power through a free and fair election. The leaders of the BNP and Prime Minister Khaleda Zia were determined to fulfill the commitments made by President Ziaur Rahman (Prime Minister Khaleda Zia's husband) to establish the Office of the Ombudsman and to make this a priority. Unfortunately, the BNP did not bring the Act into force (Halim, 2006, p. 326). The Awami League and BNP Government came to power in the seventh and eight parliamentary elections, respectively. The present ruling party, Awami League, came to power in the ninth parliamentary election and, more interestingly, they have come to power again in the tenth parliamentary election held on January 5, 2014. This was achieved without the participation of the major opposition political parties, the BNP-led coalition. However, to date, no government has taken the initiative for the establishment of the Office of the Ombudsman.\nThe Ombudsman Act 1980: a critique: The Ombudsman Act,1980 (1980) has several shortcomings and loopholes that hinder the establishment of the rule of law within the country.\nThe Office of the Ombudsman and government resistance: Strong political leadership is a very significant factor in the overall development of a country. When a political party honestly want to and have the will to, when they come to power they can reform and develop the public administration of their country. Considering this, it is unfortunate and alarming that no serious attempt was made to reform the administrative system of Bangladesh by the politicians since the Bangladesh Liberation War of 1971 (Khan, 1998, p. 173). During each election campaign, most of the political parties emphasized the need for the reformation of the public administration of Bangladesh. However, after gaining power, they did not prioritize this need for reform and they turned their attention away from the people. The parties spend a lot of money during election campaigns and even provide money to voters to secure their votes, as most of the citizens are illiterate and poor. The parties assume that they can replace this money by any illegal means during their tenure, as politics is considered a profitable business or money-making machine in Bangladesh. Hence, the majority of the MPs have business backgrounds. Evidencing this state of affairs in the Global Corruption Barometer 2013 report conducted by Hardoon and Heinrich (2013), 93 per cent of the respondents (citizens of Bangladesh) believe that police and political parties are the most corrupt institutions in Bangladesh, followed by the judiciary, parliament and government administration. The Bertelsmann Foundation (2012) Bangladesh country report pointed out that the Bangladeshi administration is burdensome and ineffective due to corruption, politicization and lack of resources. In fact, it is very difficult to find any governmental sector of Bangladesh without corruption.\nWhat next in Bangladesh for ombudsman: Lessons from the UK: England has a long history of establishing ombudsman offices. In the UK, a committee known as the Whyatt Committee (Jurists Justice, 1961) was established in 1961 for receiving complaints only from members of the House of Lords and House of Commons. Later, the jurisdiction of the committee was extended to allow the public to have direct access to the committee. The committee handled complaints against maladministration by the government and other public institutions where no courts or tribunals were available for dealing with such complaints. The concept and functions of this committee were similar to the present ombudsman institution.\nConcluding remarks: The Constitution of the People's Republic of Bangladesh 1972 includes inspiring provisions for the establishment of the rule of law and justice in the country. One of the most important provisions (Article 77), and crucial for good governance, is the establishment of the Office of the Ombudsman. This study argues that the establishment of the Office of the Ombudsman is essential in order for Bangladeshi citizens to resolve any administrative grievances made against the public administration. Hence, this study recommends that the government must establish an independent Office of the Ombudsman without further delay. The establishment of this office will ensure the protection of fundamental human rights and promote social justice and equity.\n",
"role": "user"
},
{
"content": "Purpose: It is recognized worldwide that an ombudsman system makes the public administration more transparent and accountable to the public. This paper aims to examine the provisions of the Ombudsman Act 1980, as well as the position, role and necessity of the Office of the Ombudsman in Bangladesh. It also evaluates how the ombudsman institution can act as a gateway for citizens of Bangladesh to resolve complaints against the maladministration of public administration in the country.\nDesign/methodology/approach: The study is basically qualitative in nature where both primary and secondary sources have been used. As well, a combination of analytical methods and current legislative methods, together with future legislative techniques, was used in the study.\nFindings: This study finds that the ombudsman is a vital institution for Bangladesh to eliminate maladministration, nepotism and abuse of human rights, as well as abuse of the power of the public administration. Going forward, Bangladesh needs to amend the existing Ombudsman Act 1980 and then take proper steps to firmly establish the Ombudsman Office to ensure and increase public confidence, operational effectiveness and good governance and human rights throughout the country.\nResearch limitations/implications: The main implication of this study is that it will play an important role for the development of the rule of law and human rights in Bangladesh. This study will make its readers and particularly the citizens of Bangladesh aware of the importance of the \"Office of the Ombudsman\" in Bangladesh and the existing loopholes in the current Ombudsman Act 1980. This research also provides a new avenue for scholars to contribute their knowledge and wisdom toward nation-building by further researching the Office of the Ombudsman in Bangladesh. In this way, scholars in this field can share their experiences of the role of the ombudsman to a wider audience.\nPractical implications: The study will facilitate policymakers and the government to enact an effective new law or to amend the existing law relating to the ombudsman.\nOriginality/value: The paper sets out the proposed amendment to the Ombudsman Act 1980. Hence, it will be of interest to policymakers, government, organizations of civil society and those developing countries that have not taken steps toward forming an ombudsman institution.\n",
"role": "assistant"
}
] |
Innovative behaviour of knowledge workers and social exchange attributes of financial incentive: implications for knowledge management
|
[
"Knowledge workers",
"Knowledge management",
"Social cognition",
"Innovative behaviour",
"Pay-for-performance policy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In the hypercompetitive business environment, persistent pursuit of innovation to occupy vantage positions is a critical success factor for different industrial sectors (Almatrooshi et al., 2016; Bruggemann and Meub, 2017; Hervas-Oliver and Sempere-Ripoll, 2015; Planko et al., 2017). Innovation effectiveness hinges much on the contribution of knowledge workers, who pave the way to incubate innovation and convert innovation into commercialised outputs (Aslam et al., 2017; Andreeva et al., 2017; Heisig et al., 2016; Meissner and Carayannis, 2017; van Knippenberg, 2017). Enhancing the innovative behaviour of knowledge workers, both theoretically and practically, is a main task of knowledge management (Andreeva et al., 2017; Byukusenge et al., 2016; Collins and Kehoe, 2017; Figueiredo et al., 2016; Liu and Li, 2017; Standing et al., 2016).\n2. Theoretical background: 2.1 Proposed conceptual framework\n3. Research method: 3.1 Context of research\n4. Findings: 4.1 Confirming measurement model\n5. Discussion: 5.1 Recapping main research results\n",
"role": "user"
},
{
"content": "Purpose: Enhancing the innovative behaviour of knowledge workers is a main task in knowledge management. The pay-for-performance policy is one of the management practices for innovative behaviour enhancement and has been gaining popularity in the knowledge-intensive context. However, it is still uncertain whether such practice really enhances the innovative behaviour of knowledge workers. To address this issue, this paper aims to propose and verify a conceptual framework incorporating kernel notions of social exchange, psychological empowerment and work engagement rooted in the social cognition paradigm.\nDesign/methodology/approach: The current study conducts a survey on 608 knowledge workers and their supervisors, validating the model structure and causal path pattern of the proposed framework. The causality is delineated from social exchange attributes of financial incentive, psychological empowerment and work engagement to innovative behaviour of knowledge workers.\nFindings: Perceived organisational support and perceived pay equity are primary antecedents of symbolic incentive meaning reflected in the financial incentive of the pay-for-performance policy. Symbolic incentive meaning comprising dimensions of relative position, control and personal importance relates positively to innovative behaviour of knowledge workers. Psychological empowerment and work engagement are partial mediators of the positive relationship.\nOriginality/value: The current study explicates why and how social exchange attributes of the financial incentive provided by the pay-for-performance policy may enhance innovative behaviour of knowledge workers. Implications are supplied to knowledge management scholars and practitioners to optimise the pay-for-performance policy for innovative behaviour enhancement.\n",
"role": "assistant"
}
] |
Three problems in logic-based knowledge representation
|
[
"Logic",
"Artificial Intelligence",
"Knowledge representation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction and context: This article is intended to give a brief non-technical overview of some of the technical research in which I have been involved in the last ten years or so in a particular area of mathematical logic and formal knowledge representation, largely at SLAIS, in collaboration with researchers at Imperial College London, the University of Cyprus, Harvard University in Massachusetts, and the IBM T.J. Watson Research Center in New York.The term \"knowledge representation\" is used to refer to very different things in different academic contexts. In the field of artificial intelligence (AI), usually viewed as a sub-field of computer science, it has come to mean the use of mathematical logic to produce formal models of mental structures of information and associated inference processes, with the longer term aim of using such models as specifications for machines able to simulate aspects of intelligent reasoning. Within this endeavour, a large amount of research effort has been spent on attempting to formally represent \"commonsense\" knowledge and reasoning - in other words, attempting to model the information and reasoning processes we use to go about our daily lives in our everyday environment. This has proved very difficult. It turns out, for example, to be much easier to specify how a machine could simulate reasoning about playing chess, with its well-defined rules, strategies and variables, than to specify a comprehensive, flexible and adaptive routine for making a cup of tea (including, for example, information and general planning strategies enabling you to decide what do if you have run out of tea bags and the supermarket is closed, etc.).The idea of modelling the human reasoning process via symbolic logic can, of course, be traced back to ancient Greece, and has had various proponents throughout the ages (including, notably, the librarian, mathematician and philosopher Leibniz). But the modern revival of interest in this endeavour, spurred on by its potential practical realisation via electronic computers, is largely due to the American computer scientist and logician John McCarthy. From the 1950s to the 1980s, McCarthy wrote a number of seminal papers laying out a research agenda for logic-based AI, starting with a 1959 article called \"Programs with common sense\" (McCarthy, 1959). One of McCarthy's central themes has been general reasoning about cause and effect, in particular via the notion of action. This is because such concepts are essential in the representation of key human reasoning processes such as planning. In 1969 McCarthy and Hayes (1969) proposed a logical ontology and schema for representation of such knowledge called the Situation Calculus, so called because its key ontological commitment is to view the world as progressing discretely from situation to situation via the deliberate execution of actions by an agent or agents. Initial work on the Situation Calculus has since broadened into the AI sub-field of Reasoning about Action, and it is into this area that the research described in the rest of this article fits.\n2. Event Calculus and the frame problem: 2.1. Classical Event Calculus\n3. Modular-E and the Ramification Problem: The Event Calculus is a specialised theory written in standard, general purpose logic. A complementary methodology, now widely used in the area of reasoning about actions, is to develop and analyse highly specialised logics called action languages (Gelfond and Lifschitz, 1998). Action languages retain the fundamental logical notions of models and entailment, but have a restricted syntax that allows the particular concepts and principles under investigation to be hard-wired into the semantics (i.e. the underlying definition of what a model is), rather than represented as axioms. Such an approach can be a useful extra step when conducting an in depth investigation of a particular issue, because it allows representational details that might otherwise obscure the central research questions to be abstracted away. The resultant logics can also be very useful for direct use as specifications for computer programs.Tony Kakas (at the University of Cyprus), Loizos Michael (at Harvard) and I have recently developed a particular action language called Modular-E (ME) to help us study two problems intimately related to the Frame Problem (Kakas et al., 2005). The first of these is the Ramification Problem. This is the problem of adequately representing and inferring information about the knock-on and indirect effects (ramifications) that might accompany the direct effects of an action or event. Indirect effects are commonplace in the everyday world. For instance, in our \"car\" example, turning the key will cause the engine to run, which (if the car has no oil) will in turn cause it to overheat, which in turn will cause smoke to come out of the bonnet, etc. The Ramification Problem is related to the Frame Problem in that an over-zealous solution to the latter precludes an adequate solution to the former. For example, if our solution to the Frame Problem has allowed us to conclude (by default) that turning the key doesn't cause smoke to come out of the bonnet (because it is not an explicitly stated effect of that action), we cannot then admit this as a consequence of our knowledge of indirect effects.Statements called propositions are ME's equivalent to Event Calculus axioms. Here are three c-propositions (\"c\" for \"causes\"), two h-propositions (\"h\" for \"holds\") and one o-proposition (\"o\" for \"occurs\") to describe some facts about the car domain: Equation 4 Equation 5 Equation 6 Equation 7 Equation 8 Equation 9 Because of ME's hard-wired principles of fluent persistence (part of its solution to the Frame Problem), we can infer from the above that at all times before the TurnKey action at 5 there is petrol in the car, the engine is neither running nor too hot, and there is no smoke (i.e. these will be characteristics of all models of this domain). But the propositions say nothing about whether the car battery is in fact charged or if there is oil in the car, and so we obtain four models from this domain description. In one of these models the battery is charged but there is no oil in the car, and ME's solution to the Ramification Problem ensures that in this same model there is smoke coming out of the bonnet at, for example, time 6.ME has an exceptionally robust and complete solution to the Ramification Problem. It can be used, for example, to describe domains in which some actions or effects are non-deterministic, in which simultaneously executed actions have competing causal effects, in which particular outcomes are contingent on a particular micro-ordering of indirect effects, or in which \"looping\" indirect effects give rise to rapid oscillations in fluent values (e.g. to model the mechanism of an electric buzzer). Technically it achieves this flexibility by viewing all effects as micro-processes which operate in an infinitesimal time span just \"long enough\" to explore and resolve all potential causal conflicts and outcomes.Work with action languages can feed back into work with mainstream general purpose logics. Jeremy Forth (at Imperial College London) and I are currently investigating how an earlier incarnation of ME's solution to the Ramification Problem, expressed in the Language E (Kakas and Miller, 1997a, b), can be translated into Event Calculus terms. One benefit of this is that the Event Calculus will then be able to take advantage of some of the computational techniques that have been developed for E and its successor ME.\n4. The Qualification Problem, elaboration tolerance and free will: The second problem that motivated the development of ME is known as the Qualification Problem. This has two major aspects, the endogenous problem and the exogenous problem, as described below.4.1. The Endogenous Qualification Problem\n5. Concluding remarks: The intended focus of this article is on fundamental and technical research issues rather than applications. However, the Event Calculus and its action language counterparts in various forms have been applied in a number of areas of practical information technology, such as scheduling systems, databases, accident report processing and automated legal reasoning (see the introductory section of Miller and Shanahan (2002) for a comprehensive set of references). Knowledge-based systems such as these have a potential advantage over traditional computer applications because of the flexibility and transparency of their computation (e.g. allowing reasoning both from cause to effect and vice versa, or enabling output to be accompanied by an explanation of the underlying information processing). For example, Alessandra Russo (at Imperial College London), myself and others have employed an Event Calculus implementation for use in software system specification (Russo et al., 2002), developing a tool able to indicate reasons why a software prototype does not meet particular requirements in its specification. The next stage of this work will be to take the more complete argumentation-based computational procedures for E and ME (Kakas et al., 1999, 2001) developed in collaboration with Tony Kakas and Francesca Toni (also at Imperial) and apply them in the same context, to give us an even more versatile tool for software requirements engineering.More generally, research in Reasoning about Actions is proving useful in a number of cutting-edge applications. To give just two examples, Sheila McIlraith (at Toronto) and her research team have explored how the Situation Calculus can be integrated with the DAML markup language to provide semantic web services (see, e.g. McIlraith et al., 2001), and Chitta Baral (at Arizona State University) and his team have applied the action language A to research in molecular biology, producing a system able to reason about interaction networks in cells, in particular to advance knowledge of the causes of certain types of cancer (Tran et al., 2005).These immediate benefits in contemporary computer applications and applied research are all evidence for the fundamental and over-arching reason why the AI sub-field of Knowledge Representation, exemplified by the research described in this article, should be of interest to computer scientists, information scientists, information managers and those working in related fields. This is that formal and detailed models of how information is organised and used inside our heads must be relevant to decisions about how we organise and use information more generally.\n6. Further reading: I hope that this article has prompted at least a few readers to find out more about AI, logic-based Knowledge Representation, and perhaps even the Event Calculus and ME. Russell and Norvig's (1995) book Artificial Intelligence, A Modern Approach is a particularly good general introductory text on AI. For a more in-depth but still introductory treatment of logic and Knowledge Representation, see Genesereth and Nilsson's (1987) Logical Foundations of Artificial Intelligence. Still more specifically, Davis's (1990) Representations of Commonsense Knowledge gives an excellent overview of techniques and topics in Commonsense Reasoning. It is always worth revisiting the influential and inspiring papers of McCarthy, and a fine selection of these, spanning three decades and including the three specific McCarthy papers cited in this article, is gathered together in Lifschitz (1990). Another more contemporary collection of influential research papers is contained in Hobbs and Moore (1985), including Pat Hayes's classic The Second Naive Physics Manifesto. Finally, Shanahan's Solving the Frame Problem: A Mathematical Investigation of the Common Sense Law of Inertia gives a detailed account of the development of the Event Calculus.There is, of course, also a vigorous debate at a more philosophical level about the futility or utility of AI, and about the worth of computational and logic-based models of human reasoning. Famous doubters include John Searle, Noam Chomsky and, more recently, the physicist Penrose (1989) in his book The Emperor's New Mind. A good summary of these philosophical standpoints, and an arguably effective demolition of them from a strong Darwinian perspective, is given in Dennett's (1995) Darwin's Dangerous Idea.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to give a non-technical overview of some of the technical progress made recently on tackling three fundamental problems in the area of formal knowledge representation/artificial intelligence. These are the Frame Problem, the Ramification Problem, and the Qualification Problem. The article aims to describe the development of two logic-based languages, the Event Calculus and Modular-E, to address various aspects of these issues. The article also aims to set this work in the wider context of contemporary developments in applied logic, non-monotonic reasoning and formal theories of common sense.\nDesign/methodology/approach: - The study applies symbolic logic to model aspects of human knowledge and reasoning.\nFindings: - The article finds that there are fundamental interdependencies between the three problems mentioned above. The conceptual framework shared by the Event Calculus and Modular-E is appropriate for providing principled solutions to them.\nOriginality/value: - This article provides an overview of an important approach to dealing with three fundamental issues in artificial intelligence.\n",
"role": "assistant"
}
] |
Outshopping in rural periodic markets: a retailing opportunity
|
[
"Rural",
"Retail",
"Markets",
"Outshopping",
"Periodic"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Marketers experience high distribution costs in geographically spread out rural market with widely dispersed population. One other challenge is that a number of small villages do not have retail stores. In this context marketers have experimented with periodic markets (weekly markets held in the larger villages) as an additional channel to reach the rural consumers. Marketers would be interested in understanding the influences affecting the prevalence of periodic markets compared to fixed location retail stores. Periodic markets are locations where people assemble in order to buy and sell goods and services. These markets usually operate in a weekly cycle with stable periodicity (Velayudhan, 2007). They serve the villages where they are located and 10 to 15 surrounding villages. Agricultural and forest products, village handicrafts and consumer goods are sold from temporary stalls by travelling traders. Generally produce is sold during the early part of day and in the later part of the day consumer goods are purchased and leisure activities like drinking rice beer is common (Velayudhan, 2007).\nShopping centre research: Literature on shopping centre examines the development of retail locations based on shopper and retailer behaviour. These studies have their roots in the central place theory and retail agglomeration theory. The Central place theory suggest distance and population (mass) as influences on the ability of a location to attract outshoppers (Adamchak et al., 1999; Lee and Pace, 2005; Mejia and Benjamin, 2002) while minimum differentiation between retailers explains retail agglomeration (Eppli and Benjamin, 1994). Some studies consider the assumption of central place theory too simplistic (South and Boots, 1999) and restrictive (Beguin, 1992). Review of literature on shopping centre sales suggests spatial and non-spatial factors as influencing retail sales in a location. Spatial factors include dispersion of buyers and sellers, size and mix of retailers. The dispersion pattern of buyers and sellers around a location determines the consumption benefits offered to consumers by the locations and also the demand and production costs for the sellers. The accessibility and visibility of a site as also the size and mix of retailers influence its development (Brown, 1989; Mejia and Benjamin, 2002). The non-spatial factors include retail image and mix. Multiattribute shopping is influenced by retail agglomeration. Retail agglomeration attracts even single-trip purpose possibly because of leisure and other activities, presence of pleasant shopping environment (Arentze et al., 2005). One stop shopping is sought by shoppers in large retail formats (Carpenter, 2008). In respect of retail mix the periodic market in rural location offer better variety than local shops and is one of the reasons for preferring them (Kashyap, 1998). Periodic markets attract consumers not only variety but also leisure (Velayudhan, 2007). Distance as an influence is possibly underestimated in many studies identifying a retail location (Lee and Pace, 2005). The review suggests the need to examine the influence of population and of accessibility on central retail locations in the rural areas.\nLiterature on outshopping: There are a few studies on outshopping behaviour in rural locations and in recent times there is interest in understanding inshopping behaviour in rural areas. Characteristics of consumers and benefit offerings of retailers are the influences on outshopping identified in literature.\nRetail format competition: The literature on competition between different retail formats is examined to understand the competition between periodic markets and rural retail stores. The extant literature is in the context of large chains, convenience stores and planned shopping areas and malls (Gonzalez-Benito et al., 2005; Chiou, 2009). These studies are not conclusive. It is observed that competition is within format and not between retail formats (Miller et al., 1999; Fox et al., 2004; Gonzalez-Benito et al., 2005). There is also evidence for competition between retail format as smaller retail stores are affected by larger stores (Lee et al., 2008). It is also observed that retail formats compete on different considerations (Lee et al., 2006). Consumers prefer shopping at different retail formats for different reasons (Bhatnagar and Ratchford, 2004) they may also use different methods for evaluating different formats (Anand and Sinha, 2009). Shopping orientations of consumers influence choice of retail formats (Carpenter and Brosdahl, 2011). Mixed results are also indicated with presence of cross-shopping and at the same time consumer characteristics influencing choice of retail format (Prasad and Aryasri, 2011). Consumer uses few important considerations in choice of retail stores and so new retail formats may have limited advantage in evolving markets (Sinha and Banerjee, 2004). This study proposes to examine the competition between retail formats in the rural context. The available literature on influences affecting inshopping behaviour in the rural areas (Home, 2002; Johnson et al., 2006; Miller and Kean, 1997; Miller et al., 1999; Mullis and Kim, 2011) is limited to the context of increasing competition from large retail chains in urban locations and competition from internet and catalog shopping.\nDevelopment of hypotheses: The research objectives required examining relevant literature on retail location, outshopping and competition between retail formats. The expectations developed based on literature survey are framed as hypotheses to be tested.\nMethod: Cross-sectional study is used to understand the performance of periodic markets. To study competition between periodic markets and retail stores requires understanding performance of these formats over time. The stage of retail development (stage identified by number of retail stores for a unit of population) is used as surrogate for time.\nAnalysis: The influence of access and population on the performance of periodic markets is examined through normal regression using all the variables at the same time. The population variable is entered as an additional variable to test the second hypothesis.\nDiscussion, implications and suggestions for future research: Discussion\n",
"role": "user"
},
{
"content": "Purpose: The study seeks to understand the influences on the prevalence of rural retailing institution of periodic markets. It examines the influence of access to and population of the location on the performance of periodic markets and their effect on the competition between periodic markets and local rural retail stores.\nDesign/methodology/approach: Simultaneous cross-sectional study on census data is used to understand changes in performance of periodic markets in the context of growth in retail stores.\nFindings: The superior performance of local retail shop compared to outshopping in periodic markets is contrary to observation in literature developed in the context of urban outshopping.\nPractical implications: Retail chains need to develop a separate program for location of outlets that serve rural areas as access influences retail performance. In rural areas marketers can use traditional periodic markets in addition to conventional retail store given the limited inter-format retail competition.\nOriginality/value: This is one of the few studies to explicitly examine periodic markets as a retail format. Outshopping is studied in the context of growth in local retail stores unlike earlier studies in the urban context where the local stores exhibit reduced sales.\n",
"role": "assistant"
}
] |
Measuring the impact of inter-attitudinal conflict on consumer evaluations of foreign products
|
[
"International marketing",
"Ethnocentrism",
"Country of origin",
"Thailand",
"Consumer behaviour"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: An executive summary for managers and executive readers can be found at the end of this article.\n1. Introduction: Modern consumers face a dilemma: they might prefer to buy locally made products to support their local industries, yet they also appreciate the superior quality, price, or brand image of foreign products. Their evaluations of foreign products may be shaped by their general attitudes (e.g. consumer ethnocentrism), country-specific attitudes (positive or negative), or brand-specific attitudes (e.g. special liking for a particular brand). Researchers investigate evaluations of foreign products that reflect consumer ethnocentrism (CET) (Shimp and Sharma, 1987; Swift, 1989), country-of-origin effects (Bandyopadhyay and Banerjee, 2002; Peterson and Jolibert, 1995), or animosity or affinity toward a country (Ettenson and Klein, 2005; Oberecker et al., 2008) and thus offer valuable insights into various determinants of consumers' evaluations of foreign products, yet relatively few studies investigate their combined effects (Chattalas et al., 2008; Chryssochoidis et al., 2007; Hamin and Elliott, 2006).The joint influence of these attitudes on consumer's foreign product evaluations is critical for both global and local brands: global brand marketers need to know when and how to leverage positive associations with a brand's home country, and local brand managers can benefit from knowing when and how to capitalize on CET. Furthermore, it is worthwhile to understand the relative strengths of these attitudes as a means to develop an effective branding strategy. If the country-of-origin (COO) effect of a foreign brand is positive and stronger than the influence of CET, the foreign brand should associate with its source country to accentuate brand quality perceptions. However, if the COO effect is negative and CET is strong, managers should dissociate a brand from its home country. In a joint venture with a strong local company, retaining the local brand name might be more appropriate.We investigate the combined influence of two sets of attitudes (general and country-specific) on consumers' evaluations of foreign products, using the following organization: We begin by reviewing relevant literature on general and country-specific attitudes and develop our hypotheses. Next, we describe our methodology, followed by the analysis and interpretation of results. Finally, we conclude by outlining the limitations of the study and the scope of further research in this field.\n2. Literature review: Many researchers have studied various determinants of consumers' evaluations of foreign products. Rice and Wongtada (2007) suggest a conceptual model to combine these attitudinal influences, in which consumers' evaluations of foreign products depend on the combination of general attitudes (e.g. CET); country-specific attitudes, which may be positive (affinity) or negative (animosity); and brand-specific attitudes (e.g. special liking for a particular brand).2.1 General attitudes\n3. Methodology: We investigate the relative impact of consumer ethnocentrism and consumers' attitudes toward a country on their evaluation of that country's products, as well as their intention to buy products from that country, and determine if these influences vary according to the level of risk involved. We manipulate the level of economic risk (low, medium, and high) with three product categories, selected through a pretest in which we asked 50 Thai respondents to rate a set of products on a seven-point Likert scale (7=economically \"very risky,\" 1=economically \"not at all risky\"). These respondents rated cars as the most risky and pens as the least risky product categories, and the average rating for radios comes closest to the middle of the scale. Accordingly, we use cars, pens, and radios to represent product categories with high, medium, and low levels of economic risk, respectively.3.1 Consumer ethnocentrism\n4. Results: 4.1 Ethnocentrism of Thai consumers\n5. Discussion: We find overall support for all our hypotheses except H5b, which leads us to assert that in general, consumers' attitudes toward a country have a significant impact on how they evaluate products from that country. This influence is less prominent for consumers' intentions to buy products than for judgments about product quality. These results align with findings from prior studies (e.g. Ettenson and Klein, 2005; Samiee, 1994). We also find similar results for consumer ethnocentrism (CET) and again support prior research (e.g. Ahmed and d'Astous, 2007; Wang and Chen, 2004). Finally, we find support for our hypothesis about the relative strength of country-specific attitudes and CET, in that the impact of CSA is stronger than that of CET for product judgments and intentions to buy.The interactive influence of CSA and CET on product judgments and intention to buy also is significant, such that the impact of CSA varies at different levels of CET, and vice versa. We observe somewhat contradictory results for low versus medium levels of CET in the case of weak (or negative) CSA for radio though (Figure 3), which requires some further investigation. Consumers with weak attitudes toward a country and low to medium CET appear somewhat confused about their attitudes toward products associated with that country. Our results also support the conceptual model proposed by Rice and Wongtada (2007), which postulates that consumers experience a high level of conflict when they confront a high level of CET and positive attitude (affinity), or a low level of CET and negative attitude (animosity). In the first condition, the foreign brand enjoys a positive country association but faces a strong CET, so the brand should develop a hybrid brand, in collaboration with a local brand. For example, Honda and Kawasaki sell \"Hero-Honda\" and \"Kawasaki-Bajaj\" brands of two-wheelers in India, in collaboration with two strong local brands, namely, Hero and Bajaj. In the second condition though, foreign brands would do well to limit the association with their home country and instead accentuate the quality and performance of their brand. They might attempt to \"localize\" the brand by associating it with local cultural events and projecting a good corporate citizen image. For example, multinational drug companies sponsor health clinics (e.g. free immunizations, eye surgeries) in many developing countries.The significant and negative influence of the interaction effect of ATT_US and PRODTYPE indicates that for products with lower perceived economic risk, foreign brands need to leverage their positive country image or minimize the influence of their negative country image. The significant and positive interaction effects of PRODTYPE and CET on both JGMT_US_PROD and BUY_US_PROD also demonstrate that the impact of CET increases when the perceived economic risk is high. Although we do not test a formal hypothesis, we plan to explore this relationship in more detail in the future.\n6. Conclusions: The substantial conceptual and managerial relevance of our study suggests extensions in several directions. First, our approach should be applied to many other countries, including other ASEAN countries, initially. If similar results emerge from these studies, global brands could use a regional branding strategy for all ASEAN countries. Conversely, if CET and CSA differ across individual countries, marketing executives may do well to localize parts of their global brand's strategy, especially in situations in which the country association may have an adverse effect on brand evaluations.Second, similar studies might progress into other emerging markets such as Brazil, Russia, India, and China - countries that offer perhaps the most promising growth markets in the world. Global brand managers would greatly appreciate knowledge about the direct and combined influence of general and country-specific attitudes.Third, our study contains several limitations that suggest additional research opportunities. As we noted, the data come from only one country, so replications of our study in different countries should determine whether our results are generalizable. We also used a unidimensional scale to measure affinity (positive attitude) and animosity (negative attitude) toward a country, even though Oberecker et al. (2008) argue that consumers' country-specific affinity and animosity may constitute distinct constructs (see also Wongtada et al., 2008). Therefore, a logical extension would be to examine if the combined effects of consumer ethnocentrism and country-specific affinity are distinct from the combined effect of CET with country-specific animosity.\n7. Managerial Implications: Findings from our study can help international marketers in understanding how foreign consumers differ from local consumers. Consumers in most countries prefer to buy locally made products to support their local industries, yet they also appreciate the superior quality, price, or brand image of foreign products. Their evaluations of foreign products may be shaped by their general attitudes (e.g. consumer ethnocentrism), country-specific attitudes (positive or negative), or brand-specific attitudes (e.g. special liking for a particular brand).The joint influence of these attitudes on consumer's foreign product evaluations is critical for both global and local brands: global brand marketers need to know when and how to leverage positive associations with a brand's home country, and local brand managers can benefit from knowing when and how to capitalize on CET. From consumers' point of view, they can face four combinations of influences:1. A high level of CET and negative attitude (animosity or negative COO).2. A low level of CET and positive attitude (affinity or positive COO).3. A high level of CET and positive attitude.4. A low level of CET and negative attitude.In the first condition, managers should dissociate a brand from its home country. In a joint venture with a strong local company, retaining the local brand name might be more appropriate. In the second situation, global brands should accentuate their global brand image. This is a popular branding strategy followed by many global brands (e.g. McDonald's, Coke, Pepsi) in developing countries.While the branding strategies are relatively straightforward to target consumers experiencing the first two conditions, they are not so for the third and fourth conditions. In these two conditions, consumers face a high level of conflict. In the third condition, the foreign brand enjoys a positive country association but faces a strong CET, so the brand should develop a hybrid brand, in collaboration with a local brand. For example, Honda and Kawasaki sell \"Hero-Honda\" and \"Kawasaki-Bajaj\" brands of two-wheelers in India, in collaboration with two strong local brands, namely, Hero and Bajaj respectively. In the fourth condition though, foreign brands would do well to limit the association with their home country and instead accentuate the quality and performance of their brand. They might attempt to \"localize\" the brand by associating it with local cultural events and projecting a good corporate citizen image. For example, many multinational drug companies sponsor health clinics (e.g. free immunizations, eye surgeries) in many developing countries.\n",
"role": "user"
},
{
"content": "Purpose: - Most consumers can buy products from various countries, including their own. Some prefer local products; others prefer the superior quality, price, or image of foreign products. This study aims to investigate the strength of these preferences and their effects on consumers' evaluations of, and intentions to buy, foreign products.\nDesign/methodology/approach: - With a sample of 571 Thai consumers, this study measures consumer ethnocentrism (CET), a general attitude, and country-specific attitudes toward three product categories (cars, radios, and pens) with American brand associations.\nFindings: - Thai consumers' evaluations of US products vary at different levels of consumer ethnocentrism and country-specific attitudes.\nResearch limitations/implications: - Although it extends existing research into a less developed country setting, this study still relies on data from a single country.\nPractical implications: - Managers of both local and foreign brands can make use of these findings to position their offerings appropriately in Thailand.\nOriginality/value: - This study extends the use of the CET concept to a less developed country and confirms prior results obtained in developed nations. In addition, it considers the joint effects of country-specific and general attitudes in combination.\n",
"role": "assistant"
}
] |
Financial incentives - a potent weapon for higher productivity
|
[
"Analytic hierarchy process (AHP)",
"Pareto analysis",
"Productivity improvement",
"Lean Six Sigma (LSS)",
"Heijunka (production levelling)",
"Toyota Production System (TPS)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: This paper presents the saga of the turnaround of a prestigious manufacturing plant of an Indian Integrated Steel Company. The plant manufactures primarily two types of forged steel products, namely, Types A and B. The plant was commissioned way back in the seventies with an annual installed rated capacity (termed as rated capacities (RC) herein after) of 90,000 of Type A and 40,000 of Type B (equivalent to 40,000 and 15,000 tonnes, respectively, in terms of weight). Raw materials for these two products are small steel ingots that are also manufactured internally in other plants of the company. Type A has a few varieties ranging from simple to complex shapes; the latter requiring extra precision machining; while Type B is comparatively simpler and has no variety. Unfortunately, however, performance and yield of this plant, both in respect of quantity and quality, has been a major problem since its inception. Year after year, productivity of this unit remained stagnant around 60-65 percent RC, and over time it became a major challenge for the company. The problem was more painful because all other adjacent manufacturing units, also commissioned during the same time, were performing well - some even well above their RC levels. The management carried out all kinds of efforts from method/process improvements to introduction of motivational incentive plans for the employees. Although productivity improved temporarily this was not to the expected height nor was the improvement sustained for a substantial period. Management, becoming desperate, then decided to technologically upgrade the production line for Type A and to reduce production of Type B. The RC of both lines were also reassessed and downgraded to 70,000 of Type A and 10,000 of Type B (equivalent to 26,800 and 4,700 tonnes in weight, respectively). Apart from these measures, management implemented a number of other steps such as reducing annual production targets, monitoring by senior management, enhancing monetary incentives for employees, and partial outsourcing of precision machining. However, these measures also worked for an initial small period before falling back to the same old levels.\n2. Concepts: 2.1 Six Sigma\n3. Productivity improvement through financial incentives: Productivity in any manufacturing plant is adversely affected due to imbalances in one or more elements of the important four M(s): Man, Material, Machine or Method. In striving for productivity improvement, one needs to examine the reasons (sub-causes) for productivity deficiencies in all four elements and determine which affect productivity the most. This analysis problem is a chronic one, the task is more difficult and complicated than people often appreciate. Over time, in the manufacturing plant under study, many people have tried their best to improve one or more of these causes. But in the absence of a structured approach, most of those efforts were random and, hence, failed. It was therefore important in this case study to approach the problem systematically with the DMAIC methodology of LSS and, thus, identify all possible reasons for poor productivity and present them in fishbone diagrams for further analysis. Pareto's 80/20 analysis would help segregate the vital 20 percent that caused the most problems and AHP would also be useful for prioritizing these for consideration for improvement. In carrying out these activities, co-operation and participation of the shop's workmen, management and the top management of the company would also be required to be suitably integrated. Accordingly, the improvement and implementation process was planned and accomplished in two phases as described next.\n4. Conclusion: Apart from becoming a great savior for a plant which was at the brink of closure, the actions described in this case study have many other significant contributions too:\n",
"role": "user"
},
{
"content": "Purpose: This paper, in fact is the saga of turnaround of an ailing PSU plant, which in spite of all kinds of improvements measures taken up by the Company had never seen productivity beyond 65 percent of installed capacity. The purpose of this paper is also to showcase the amazing power of financial incentives in enhancing productivity, if rightly designed. On the other hand, it will also serve a lesson of caution to the users by highlighting the extent of damages what a faulty incentive plan can cause.\nDesign/methodology/approach: The methodology of Lean Six Sigma helped analyzing and improving the problem and tools like \"Fishbone diagram\" and \"Analytical Hierarchy process\" were very handy in identifying root causes for this complex problem and prioritization of those, respectively.\nFindings: Root cause of low productivity being identified as \"demotivated workforce on account of poor incentive earnings\", the existing financial incentive plans were given a relook. LSS tools like SIPOC, \"heijunka\", \"brainstorming\" etc. were applied for revealing critical faults in the existing financial incentive schemes. Some unorthodox and very common methods were adopted in modifications and implementation of incentive plans.\nResearch limitations/implications: Modification of incentive scheme involving labor union bargain is commonly resisted by both the parties, i.e. labor unions as well as the management. Although their interest behind the same remains different. One fears to loose, while other is afraid of conceding more. This case study was not an exception too.\nPractical implications: Expecting resistances, a good and thorough Shadow working with all kinds of \"extremities tests\" were prepared. This along with complete transparency followed by well explanations made both the parties happy. Accordingly, the modified incentive plans were agreed upon and subsequently were approved by the management for implementation. Few other remedies and countermeasures suggested were also implemented.\nSocial implications: The entire workforce was extremely happy and highly motivated. Provisions of equal incentive weightage with ample individual scope of earnings for both rival production groups in the modified incentive scheme successfully converted the inter-group hatred into healthy completion. Both the groups were gearing for much higher performance and earn more. Self-motivations were turned into group motivation, which is always a blessings for any incentive scheme.\nOriginality/value: Post-implementation period results were extra ordinary and unprecedented. Productivity was significantly enhanced to 15 percent in first six months, which increase up to 39 percent next year. Customer order and quality fulfillment met for the first time, relieving the management from great embarrassment. The annual incremental financial gain was more than Rs 1,000 millions. The methodology of identification of the root causes and the unique style of finding the solution are original in nature and would be helpful and guide for students, professionals of financial incentive designers, industrial engineers, managers and entrepreneurs.\n",
"role": "assistant"
}
] |
The applicability of the Viable Systems Model as a diagnostic for small to medium sized enterprises
|
[
"Business performance",
"Business analysis",
"Business improvement",
"Performance management",
"Small to medium‐sized enterprises"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The foundations of management cybernetics is to be found in the work of Wiener (1948) who discussed the science of communication and control in the animal and the machine. This work was built upon substantially by Stafford Beer in his work over many decades but which has its origins in Beer (1959). The author proposed the viable system model (VSM), an holistic approach to assessing organisational viability and identifying the root cause of any problems that might pose a threat to that viability in organisations. The VSM has been applied to a wide range of situations, however, the application of the VSM to organisations is one of the key applications of the model.The VSM was partly intended as a guide for any goal seeking, adaptive system (Beer, 1981, p. 5). In particular, for making recommendations that might improve: business performance, organisational structure, information flows, organisational identity, core mission and purpose, coordination and control, monitoring and intelligence, and internal variety mechanisms to attenuate, amplify and absorb external variety. The aim of the model was to take the focus away from hierarchy and structure toward decentralisation of control processes and critical linkages by providing a vehicle for identifying dyadic relationships.Advocates of the VSM claim a general applicability for the model (Jackson, 2004, p. 560). The same author outlines the advantages of the theory. First, it is a rich representation of the organisation and insightful in its portrayal of relations with the environment. Second, it provides a role and scientific justification for empowerment and democracy in the organisation by promoting the role of the parts of the system and the workforce in the overall purposes pursued. Third, its recursive nature allows it to cope with organisations whose parts are both vertically and horizontally interdependent. Finally, the model enables a focus on the important sources of command and control and it is a suitable starting point for information systems design. Although research demonstrating successful application of the VSM to organisation contexts exists, Whittaker (2003) argues that the work of Beer is \"scandalously undervalued\". Brocklesby and Cummings (1996) identify a lack of publication accessible to the practitioner community as being a primary reason for this absence. Some attempts have been made to make the work accessible to the practitioner community; for example Diagnosing the System for Organizations (Beer, 1985). However, a key criticism of this work is that it introduces the VSM and asks the reader to situate her own organisation on the model as opposed to presenting the reader with a more practice-oriented approach. Application of the VSM in small- to medium-sized enterprises (SMEs) is substantially more limited.This paper serves aims to present an approach to using the VSM as a diagnostic tool for SMEs to undertake an audit of viability. The paper presents a diagnostic questionnaire that maps on the VSM and which was administered to a number of SMEs; this paper aims to demonstrate the outcome of using the tool in various case-study organisations.The paper first introduces the VSM through a review of the literature, using the literature to define a gap in current research. The next section defines how the research explores this gap, explaining both the methodology that underpins the research and also how the VSM model was used in this research. The main findings of the research are then discussed through the analysis of empirical data gathered through fieldwork. These results are then discussed before the paper provides some conclusions regarding the usefulness of the VSM to diagnose issues with viability in SMEs.\n2. Literature review: This section summarises and critiques the literature on the VSM.2.1 The VSM\n3. Research design: This section first describes the research methodology before documenting how the VSM model was used to identify the \"target system\". This section also documents the diagnostic interview used to identify issues with viability.3.1 Methodology\n4. Findings: The research suggests that the VSM has a number of advantages for the study of viability in SMEs. Some of the main findings are discussed with using cases from the research.The most common operational failing for the small businesses involved was the absence of sub-systems for control, coordination and communication. These are systemic mechanisms found at all levels in the VSM as self-sustaining feedback loops and vital for proper links between operations and other parts of the organisation (Table I).\n5. Discussion: This research set out to identify whether the VSM is an appropriate tool to diagnose viability in SMEs.One of the major challenges in using the VSM is ascribing a primary operational purpose (e.g. to provide goods and/or services) to the organisation which in VSM terms is called the SiF. This step was a beneficial exercise as thinking carefully about the system and its operational purposes brings new issues and insights into consideration concerning the real purpose of the organisations. Whilst it was felt that for a small business the task should be easy since there are only to few people involved lowering the risk that the SiF is open to a multitude of subjective interpretations. In reality, often the firmly entrenched views of an owner or proprietor are hard to influence. Nevertheless, the fundamental objectives of each organisation must be confirmed in order to delineate the external and internal environments.One of the key findings from the research was that the method was successful in identifying issues with viability. Table II summarises the analysis of the case data.On reflecting on why the approach was successful, it is necessary to return to the holistic nature of the VSM model which requires any intervention to view the organisation in its entirety with an intricate analysis of detail recognising cause and effect between sub-systems. Whilst the issues presented themselves as being operational problems the VSM diagnostic model viewed the operations as a part of a wider socio-technical system. This view places attention on the many systemic components and the linked cause and effect implications of change to business interfaces that are involved in various operational processes and designed to encompass entire supply networks.It was found that organisational activities were often instinctive and unconscious responses to internal and external \"variety\". Internal homoeostasis was not present in Companies A, B and D whilst external homeostasis was not present in Companies C, E, F, G and H. This meant that variety was not responded to and where this was the case the viability of these organisations was extremely weak. \"Requisite variety\" is a fundamental theorem underpinning the VSM and refers to an organisation's capability to respond appropriately to its environment.It also seems clear that the VSM offers many advantages for organisational research. The structured and holistic viewpoint it encourages in seeing an organisation responding to its internal challenges and external environment, are perhaps of most note in this respect.Despite the importance of establishing the SiF and thus the organisation's raison d'etre some very small and new businesses clearly struggle with overall direction due to a lack of future planning (System 4), whether informal or formal - for example in the case of Company G. Intervention did help long-term thinking and some external reflection.It perhaps comes as no real surprise that many small businesses become entwined in the day-to-day running of all aspects of the business, often at the expense of doing any one aspect properly. Symptoms such as lost sales, inefficiencies and high levels of waste were symptoms often seen in the research.The VSM proved an invaluable tool for targeting the root of the problem in order to highlight remedial action needed. Analysing an SME using the VSM provided insights into the intricate details of an organisation.For viability the system needs to match the requisite variety of the external environment. Viability is maintained by engaging in different activities, not necessarily reducing activities to become simpler to manage, a common misconception, but keeping activities and functions from interfering with each other, whilst managing them together, and focusing on the future, and doing so in the context of an identity within which the interests of the whole can be considered. Such a task presents diverse challenges to an SME and thus the long-term viability and hence survivability of an SME.For the small business, the research offered an important opportunity to \"step back\" from the daily minutiae and \"fire-fighting\" allowing a more reflective practice with a longer-term, external orientation: For example, two companies who were involved in the research described the process as being \"like standing in front of a mirror\" and \"a launch pad for the future\".\n6. Conclusions: The VSM has been used in this study as the foundation of a diagnostic intervention focused upon the activities of a number of small businesses. The approach used action research in an effort to aid the firms involved and, it is hoped, to develop their own mechanisms of assistance and be less reliant upon external agencies.Perhaps the main difficulty in using the VSM to diagnose small business viability is the propensity to use the model to provide a schematic representation of a logical hierarchy replicating a bureaucratic organisational structure with power vested at the top and filtering down. This is especially true for the small business having power often vested in an owner/proprietor. For the inexperienced, this could encourage a mere replication of the existing organisational chart in VSM format, thus little benefit will be yielded in consequence.However, returning to the aim of the study, the research suggests that when used appropriately the VSM is valuable for understanding operations and assessing their viability and thus for developing increased flexibility, efficiency and effectiveness. These are all operational strengths of firms of this size, but can only be properly understood at the level of the interface between operations and the task environment and between operational and other support sub-systems.For the small firm flexibility is about changing existing states quickly (speed of response). Larger firms it seems may have difficulties with speed of change in this sense but have the excess resources to potentially facilitate change into a greater number of states or the necessary \"requisite variety\". From this study it became clear that the smaller business was operationally more nimble but could not always achieve the number of states necessary to cope with the variety from its external environment. They have the advantages of velocity over variety.Both operational flexibility and agility were of equal importance to this study and the adapted VSM was particularly helpful in seeking the holistic interventions necessary across the whole organisation and its external linkages. Future research would do well to expand upon this particular aspect as it remains key for the success of many small- and medium-sized firms and correlates directly to the wealth, health and well-being of SMEs as a whole.Lastly, the positive findings of the research to date should help to promote the formulation and testing of further research into viability in SMEs. This research was conducted amongst manufacturing SMEs: additional research might involve detailed case studies and further empirical work for a broader number of SMEs.\n",
"role": "user"
},
{
"content": "Purpose: - This paper presents research into the use of the Viable Systems Model (VSM) in small to medium sized enterprises (SMEs). Research on the VSM has been focussed on large organisations. The purpose of this paper is to explore the usefulness of the VSM in diagnosing issues of viability in SMEs.\nDesign/methodology/approach: - Case study research was undertaken in which semi structured diagnostic interviews took place with SMEs using the VSM as a diagnostic tool. The aim was to investigate whether the VSM would assist in diagnosing problems with viability and whether it could then prescribe and facilitate improvements in operations. Qualitative case study evidence is presented that shows the results of these interventions.\nFindings: - The paper reports some examples of operational problems that were exposed through the intervention process and shows some general conclusions to support the use of VSM for analysing operations and supporting small business viability. A number of \"threats to viability\" themes were apparent and these are discussed.\nPractical implications: - The semi-structured interview protocol created for the research could be used to diagnose viability issues in other SMEs and thus is of practical relevance to other organisations.\nOriginality/value: - The VSM originates from the theory of cybernetics which can be broadly defined as \"the science of effective organisation\". Previous use of the model has focused upon the viability of large organisations; the applicability of the VSM in a small business operational context has received little attention. The paper demonstrates the benefits of the using the VSM to pinpoint problems that threaten SME viability.\n",
"role": "assistant"
}
] |
Multilevel analysis of the relationship between principals' perceived practices of instructional leadership and teachers' self-efficacy perceptions
|
[
"Instructional leadership",
"School management",
"Principal leadership",
"Student engagement",
"Classroom management",
"Teacher self-efficacy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Contemporary educational research has provided persuasive evidence that the teacher is the most important school-related factor to influence student learning outcomes (Aaronson et al., 2007; Coleman et al., 1966; Ferguson, 1991; Odden et al., 2004; Rockoff, 2004). It has been found that effective teachers, and thereby effective teaching, are the most indispensable component of efforts to raise the bar of average achievement, and to narrow the gap between students with distinct socio-economic backgrounds (Darling-Hammond, 2000, 2010; Odden et al., 2004). Given such substantial findings, increased international attention is being given to the quality of teaching and effective teachers, particularly in the past couple of decades, with an unprecedented emphasis on accountability.\nTheoretical foundations: The theoretical foundations of the present study begin with a discussion on the concept of self-efficacy, which is the study's main dependent variable, followed by an introduction to the concept of instructional leadership. Finally, the conceptual framework that guided the study is provided.\nConceptual framework: In an educational context, school processes have been indicated as the most important factors to bring about desired outcomes (Reynolds et al., 2002); and while principals are key personnel to exert direct influence on school processes, teachers are the ones who have a direct effect on student outcomes (Hallinger et al., 1996; Hallinger and Heck, 1996), as shown in Figure 1. According to this perspective, principals' leadership practices should play a critical role in creating effective school processes. Some studies have provided evidence that principals can affect teachers' self-efficacy perceptions by engaging in instructional leadership practices (Calik et al., 2012; Duyar et al., 2013). This study is designed to examine whether the same is true when taking into account teachers self-efficacy perceptions in different aspects of teaching and interactive processes. Specifically, the purpose of the current research is to investigate the impact of school principals' instructional leadership perceptions on teacher self-efficacy across countries (using the data available for 34 countries). A two-level mixed model was adopted to investigate the relationship between principals' perceptions of instructional leadership practice and teacher efficacy in classroom management, instruction, and student engagement, while controlling for school-level variation of principals (gender, experience, education, and professional development) and school background (type, location, size, funding source), as well as teacher-level fixed effect (gender, age, education, experience, employment status, tenure, and training).\nData and research methods: Data source\nFindings: The most important finding for this study is the significant and positive correlation between principal perceived instructional leadership and teacher self-efficacy, which has been consistent through the analysis (Table II).\nDiscussion and conclusion: The purpose of the present study was to investigate the extent to which principals' instructional leadership predicts teacher self-efficacy, in order to identify whether a relationship exists between principals' perceived instructional leadership practices and teachers' perceived self-efficacy in classroom management, instruction, and student engagement, controlling for several principal, teacher, and school characteristics. Both teacher and school-level data obtained from a large scale data set, the 2013 TALIS, which was administered by the OECD in 2013, was employed to reveal the relationship between instructional leadership and teacher self-efficacy.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the extent to which principals' instructional leadership predicts teacher self-efficacy, in order to identify whether a relationship exists between principals' perceived instructional leadership practices and teachers perceived self-efficacy in classroom management, instruction, and student engagement, while controlling for several principal, teacher, and school characteristics.\nDesign/methodology/approach: The data employed in this study were both teacher- and school-level data sets obtained from the Teaching and Learning International Survey, which was administered by the Organization for Economic Co-operation and Development in 2013. A two-level mixed model was employed in the analysis of the data by adding adjusted weights at both levels for the complex survey data.\nFindings: The results indicated a significant and positive relationship between principals' perceived instructional leadership practice and teachers' self-efficacy in all three aspects. Also, gender, experience, tenure status, and formal in-service training of teachers were found to be the key factors that have significant effects on teachers' self-efficacy perceptions.\nPractical implications: Findings suggested that international effort for mandating instructional leadership in schools is a worthwhile strategy, which can help teachers develop a greater sense of ability in classroom management, instruction, and student engagement.\nOriginality/value: This study is the first of its kind to examine the relationship between instructional leadership and teachers' perceived self-efficacy in multiple areas related to teaching.\n",
"role": "assistant"
}
] |
Are diets healthier when they contain branded foods?
|
[
"Branding",
"Costs",
"Nutrition",
"Diet",
"Prices",
"Nutrition labelling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Numerous studies have shown a positive association between food quality and food price (e.g. healthier foods are generally more expensive) (Darmon and Maillot, 2010; Drewnowski, 2010). The first studies that revealed this positive association for both single foods and food groups were based on French data (Darmon et al., 2003, 2005; Maillot et al., 2007), followed by similar findings in numerous later studies (Drewnowski, 2010; Katz et al., 2011). Generally, these studies compared the energy density (kcal/gram) and energy costs (price/kcal) of different food items and found that energy-dense products (high in fat and sugar) were relatively cheaper than more nutrient rich products such as fruits and vegetables. Moreover, at a diet level, studies showed that comparatively healthier diets are relatively more expensive than unhealthier diets (Cade et al., 1999; Drewnowski et al., 2004, 2007; Maillot et al., 2007; Waterlander et al., 2010).\nMethods: This study examined the difference in nutrient content and price of daily food intake data designed to contain food products exclusively from either LB, GB (GB) or DB. In addition, the nutrient values of these three daily intake models were evaluated against national dietary guidelines.\nResults: Available nutrient information\nDiscussion: This study examined the difference in nutrient content and price of real daily food intake data modeled to contain food products exclusively from either LB, GB or DB. Data included n=40 diets and n=430 food products. Results revealed that daily intake models designed containing exclusively discount brands were significantly cheaper but had, except for sodium, similar nutrient contents compared to daily intake models designed containing leading or GB. While there were some observable differences in other nutrient contents, these were not statistically significant and had different directions for different nutrients. These results indicate that discounts brands are both an economic and good quality option for people with limited financial resources.\nConclusion: Price is one of the most important factors in food choice and especially people with a lower socio economic status may experience barriers in selecting certain foods (Inglis et al., 2005; Steenhuis et al., 2011; Waterlander et al., 2010). The current worldwide financial crisis makes that the purchasing power of many consumers is declining and that a growing number of consumers is forced to make economic driven food choices. Dutch market trend data show that the popularity of cheap discount stores is growing and that people buy more general or home brands compared to LB (Strien and Wierenga, 2009). So far, there was limited evidence showing how these economic food choices would affect diet quality. The large and persistent social inequalities in obesity and overweight by education level and socio-economic status in OECD countries warrant close examination of economic food choices on nutrient intake and health (Devaux and Sassi, 2012).\n",
"role": "user"
},
{
"content": "Purpose: Market trend data show a growing popularity of discount food stores and of cheaper food products as opposed to more expensive leading brands (LB). Unexpectedly little is known about how these economic food choices affect diet quality and/or health. The purpose of this paper is to examine differences in nutrient content and cost of daily food intake data modeled to contain food exclusively from either LB, generic brands (GB) or discount brands (DB).\nDesign/methodology/approach: This study analyzed nutrition information of 430 food products that were selected based on dietary intake data from a sub-sample of the Amsterdam Growth and Health Longitudinal Study. Data were collected in Dutch supermarkets, where information was copied from back-of-pack nutrition tables.\nFindings: Results showed that there were no statistically significant differences between the LB, GB or DB daily intake models in energy (kJ), protein, carbohydrates, total fat, saturated fatty acids, fiber or added sugar contents. However, there were significant differences in sodium content where LB had significant less sodium compared to GB and DB. Also, there were significant price differences: LB vs GB+2.75/day; LB vs DB+7.17/day; GB vs DB+4.42/day.\nOriginality/value: To the knowledge, this is the first study analyzing differences in nutrient content and price of leading, generic and discount food brand formats on a diet level. The analysis revealed that there is little reason to suggest that dietary nutrient quality is negatively affected by an increased consumption of DB products. Indeed, the substantial price difference between leading and DB suggests that discount products provide a reasonable alternative to LB.\n",
"role": "assistant"
}
] |
Mnemonics in the Mouseion: Considerations on spatial mnemonics as a tool for classification and retrieval
|
[
"Library history",
"Library in Alexandria",
"Library in Pergamum",
"Memory art",
"Mnemonics",
"Pinakes"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Generally speaking, mnemonics was a part of classification and retrieval processes in all libraries before print. Mary Carruthers (2008) writes in The Book of Memory about this, and with great clarity demonstrates how it worked. Figures, illuminations, notes, numbers and the overall structure of books - and collections of books; libraries - constituted a mental grid, that scholars after intense mnemonic practice could browse in their memory. Carruthers (2008) focusses on memory techniques in medieval books and libraries, but she argues several times (p. 129) that the relation between them date back to antiquity. One of her many and convincing examples deals with the Canon Tables of Eusebius, a concordance added to almost all medieval bibles (see Plate 1):\nOn spatial Hellenistic mnemonics: But first, the existence of spatial mnemonics in the Hellenistic era must be ascertained. It is a question that still today divides scholarship. Accordingly, Jocelyn P. Small (1997) thoughts on the emergence of spatial mnemonics will be discussed below. In her book Wax Tablets of the Mind (1997) she presents a persuasive assumption on the reason why spatial mnemonics emerged, but she places it in the wrong era.\nOn the use of mnemonics in the Hellenistic libraries: As just mentioned, the entire architecture of Hellenistic libraries could not have been reflected in mnemonics. But a certain part of the library architecture could indeed. In the following, the architecture of the Attalid library in Pergamum is discussed as basis for spatial mnemonics.\nOn the effectiveness of mnemonics in the Hellenistic libraries: In the following the reader will get an impression of the effectiveness of such spatial mnemonics. The first feature that I would like to discuss is - I believe - the most important in all great collections of written knowledge.\nConclusion: The purpose of this paper has been to consider how mnemonics could have played a role in the libraries of the Hellenistic era, initially by introducing thorough studies proving the existence of mnemonics in libraries in antiquity. Then Small (1997) point of view, that it was only Roman mnemonics that was spatial, was argued against and refuted. Nevertheless, her idea that it was the increase in written matter that caused complex spatial mnemonics to occur was accepted. It was argued, that Hellenistic library architecture and mnemonic architecture could not have had a 1:1 relation between them, since library architecture simply changed too much and was dominated by pragmatic measures and not ideal ones, like mnemonic architecture.\n",
"role": "user"
},
{
"content": "Purpose: Mnemonics was a tool in classification and information seeking processes in pre-print libraries. The purpose of this paper is to study the role of spatial mnemonics in Hellenistic libraries, including the one in Alexandria.\nDesign/methodology/approach: Since library- and information science has not explored this subject in depth, philology, rhetoric, book-history and archeology constitute the core literature. From this literature, the role of mnemonics in the libraries is discussed.\nFindings: A new description of the practice of classification and retrieval in Hellenistic libraries, based on spatial mnemonics.\nOriginality/value: This paper is a new analysis of spatial mnemonics in the Hellenistic libraries. As will become clear, they blend easily and logically with each other.\n",
"role": "assistant"
}
] |
How important is the "social" in social networking? A perceived value empirical investigation
|
[
"Perceptions",
"Social media",
"Technology adoption",
"Social networking",
"Perceived value",
"Twitter",
"User satisfaction"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Social networking applications recorded an unprecedented success in just few of the recent years. For instance, people in the USA have been spending 22 per cent of the time they are online on social media sites while nine million users in Australia have been spending almost nine hours per month, on average, using top social media applications (Wikipedia, 2012). Despite these astonishing figures, the social networking domain is still little understood. Definitions and borders of the social networking (also called social media) phenomenon are still under debate. However, scholars seem to agree that content generated by users is the key feature of any social networking application. For instance, some conceptualization attempts define social media as \"a group of Internet-based applications that build on the ideological and technological foundations of Web 2.0, and that allow the creation and exchange of User Generated Content\" (Kaplan and Haenlein, 2010).\n2. Theoretical background: Investigating factors of adoption of new information technologies or applications has been a traditional area of research in IS. In addition to the popular models and theories validated in various studies (for a detailed review, see Venkatesh et al., 2002 study), a relatively newer approach has been to examine the adoption of an IT from a value perspective. This path was adapted from other disciplines like consumer behaviour or economics where value is used to explain why people buy some things or opt to make some expenses (Turel et al., 2010). Value is considered to source from the actual interaction with (or expected use of) a product or service and to reflect an overall perception upon their importance for an individual. Thus, following a rationale borrowed from consumer behaviour, value is captured as a perceived value concept through individual views on the difference between \"what is received and what is given\" (Zeithaml, 1988).\n3. Research model: To investigate the perceived value of Twitter, this study proposes a multi-faceted perceived value construct sourcing from consumer behaviour and IS research. This multi-dimensional value perception should have a positive influence on the intention to use the social medium since people would use a service if they perceive it as valuable for various reasons (Ho and Ko, 2008). Taking into the account of the above, the following hypothesis is proposed:\n4. Methodology: Model and hypotheses were tested through a cross-sectional experiment comprising an online survey. To ensure reliable psychometric properties, survey questions measuring the items of the latent variables were adapted from measures previously validated in consumer behaviour and IS research, as reported in top publications (Cyr et al., 2009; Kim and Han, 2009; Turel et al., 2007; Van Slyke et al., 2007; Venkatesh and Davis, 2000). Survey measures and their related constructs are presented in the Appendix.\n5. Main findings: A total of 134 valid responses were recorded at the end of the three-month data collection. A demographic analysis indicated that respondents were 39.0 years old on average, 60.8 per cent female and 39.2 per cent male. Participants reported having an average experience with Twitter of 1.6 years and checking the service 20.1 times a week, on average. A per cent of 61.2 of the respondents reported having a Twitter account, 28.4 per cent not having an account, while the rest of 10.4 per cent preferred to not answer this question. Participants having an account reported posting 10.2 messages per week, on average. They were following (i.e. subscribing to the posts of) 84.8 accounts and were having 82.7 followers, on average. All 134 participant valid responses were subjected to the data analysis.\n6. Discussion and conclusions: The objective of this paper has been to propose a value-based theoretical model to explain the adoption of Twitter, a very popular social networking application. A model based on consumer behaviour and IS literature findings as well as on theoretical reasoning was built and tested empirically with 134 North-American respondents recruited online.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to report on a value-based empirical investigation of the adoption of Twitter social networking application. The unprecedented popularity of social networking applications in a short time period warrants exploring theory-based reasons of their success.\nDesign/methodology/approach: A cross-sectional survey-based study to elicit user views on Twitter was conducted with participants recruited through the web site of a North-American university.\nFindings: All facets of perceived value considered in the study (utilitarian, hedonic and social) had a significant and relatively strong influence on consumer intent to use Twitter. Quite surprisingly for a social networking application, though, the social value facet had comparatively the weakest contribution in the use equation.\nResearch limitations/implications: User value perception might have been influenced by the features of the actual social networking application under scrutiny (i.e. Twitter in this case).\nPractical implications: To maximize the chances of success of new social networking applications, developers and marketers of these media should focus on the hedonic and utilitarian sides of their perceived value.\nSocial implications: Additional efforts are necessary to better understand the reasons and factors leading to a comparatively lower social value perception of a social networking application, compared to its hedonic and utilitarian values.\nOriginality/value: Overall, the study opens the door for investigating user perceptions on popular social networking applications in an effort to understand the unparalleled success of these services in a short time period.\n",
"role": "assistant"
}
] |
Emotional intelligence competencies in engineer's effectiveness and engagement
|
[
"Engagement",
"Effectiveness",
"Emotional intelligence",
"Shared vision"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Engineers make things. They make things better. The profession of engineering has been at the heart of the development of societies and economies throughout time. Given that engineers often create the forward movement of our societies, the education, development, and training of engineers should be of major research interest. Within the engineering field, its journals and courses, it is not surprising that the major focus remains on the acquisition and enhancement of the technical skills, knowledge and imagination believed to be needed to be effective as an engineer.\nESI competencies: Emotional intelligence is formally defined as \"the ability to monitor one's own and others' feelings and emotions, to discriminate among them and to use this information to guide one's thinking and actions,\" (Salovey and Mayer, 1990, p. 189). Some scholars, however, claim these are not forms of intelligence (Ashkanasy and Daus, 2005). Others claim that they are a capability emanating from neural activity that appears at a different level than an internal processing ability (Amdurer et al., 2015). While debates regarding the exact nature of emotional intelligence continue, what is most important in this paper is the distinction between technical skills and emotional-social skills. This distinction is relatively clear at the behavioral level. What has become apparent more recently is that ESI is also neurologically distinct from technical skills. Technical skills rely on cognitive, IQ-type abilities based in the task positive network predominantly in the neocortex. In comparison, ESI competence relies on social and emotional reasoning abilities based in the default mode network (Boyatzis et al., 2014).\nESI, relationships, and engineer effectiveness: ESI competencies have been found to predict outstanding leader, manager, or professional performance across a variety of professions (Boyatzis, 2009; Druskat et al., 2005) and have been proposed as important to engineers' capability (Passow, 2012; Lappalainen, 2015; Lopes et al., 2015; Bish et al., 2014). Synthesizing the literature, the competencies appear in three clusters: cognitive intelligence competencies, such as systems thinking or pattern recognition; emotional intelligence competencies, such as adaptability, emotional self-control, emotional self-awareness, positive outlook, and achievement orientation; and Social intelligence competencies, such as empathy, organizational awareness, inspirational leadership, influence, coaching and mentoring, conflict management (i.e. negotiation), and teamwork. The latter two clusters are used to measure ESI for this study. There are other cognitive abilities that appear to be \"threshold competencies,\" such as deductive reasoning and memory (Boyatzis, 1982). This means that they are needed to be adequate, but more use of them does not lead to effectiveness, such as knowledge (technical and functional); deductive reasoning, and quantitative reasoning (Boyatzis, 1982).\nESI, relationships, and job engagement: Whether or not people choose to use their capabilities at work has been described as their degree of engagement at work. Engagement is conceptualized as a relatively stable affective-cognitive state defined as a positive, fulfilling work-related state of mind that is characterized by vigor, dedication, and absorption (Schaufeli et al., 2006, p. 702). Engagement has become a major concern in management circles since the Gallop Report (2016) revealing that 68 percent of full-time workers in the USA do not feel engaged in their work. The belief is that without engagement, not only will people not bring their talent to work and use it, but retention of key people will be lower. Within the STEM field, job engagement has been shown to predict long-term commitment to science, technology, and mathematics careers of women professionals (Buse and Bilimoria, 2014).\nMethod: Sample\nResults: In terms of simple correlations, engineer effectiveness was significantly related to ESI competencies as seen by others, general mental ability (g) and openness. Engineer engagement was significantly correlated with personality traits of extroversion and conscientiousness, and shared vision. Job engagement was positively correlated with shared vision, extraversion, and conscientiousness. Shared vision was negatively correlated with neuroticism and positively correlated with extraversion and openness. Finally, tenure was negatively correlated with general mental ability (g) and extraversion and positively correlated with age. Age and gender were not significantly correlated with any other variable, as shown in Table I.\nDiscussion: ESI predicted engineer effectiveness to a significant and substantial degree above and beyond g, personality and quality of relationships. Therefore, H1 was supported. There was no significant relationship between shared vision and engineer effectiveness. Thus, H2 was not supported. ESI did not predict engagement above and beyond g, personality, or quality of relationships. Therefore, H3 was not supported. Engagement was shown to be predicted by the quality of relationships in terms of degree of shared vision and nothing else. Therefore, H4 was supported.\nDeveloping ESI and shared vision: The results of this study strongly suggest that some additional components addressing ESI competencies and the building of shared vision in project teams should be added to the training and education of engineers. Singh and Jha (2012) have even reported that engineering teachers and students report that several EI competencies are critical to teacher effectiveness. In this study, emotional stability, self-motivation, managing relations, self-awareness, and integrity emerged as the best predictors of teacher effectiveness from both the instructors and the students. Cognitive ability (g) was not measured in the study.\nLimitations: There are a number of limitations to this study and generalizations from its findings. Perhaps the most glaring was the sample size (n=40). The sample size was smaller than hoped because of the low-response rate. This may have resulted in more of a volunteer bias than is often encountered in survey research in organizations. Further, due to the limited sample, it may be possible that the findings may be a function of organizational culture. Although it is difficult to collect the amount of data from engineers in organizations sensitive to time demands, replication of these findings is imperative to increase confidence in the results. We believe this study provides a solid foundation and positive outlook for future research with ESI and engineers that should be encouraging to scholars.\nConclusion: The primary purpose of this paper was to show the importance of ESI and its relationship to an engineer's effectiveness and engagement and to inform the design of undergraduate and graduate programs, as well as organizational training to be more balanced in emphasizing communications, teamwork, and leadership. The findings of this research were able to show that the effectiveness of engineers is dependent, in part, on the degree of ESI behavior they demonstrate to others. Their own engagement in their work is dependent, in part, on the degree of shared vision they view in their project teams. These two findings suggest that engineer education and training should include development of ESI as well as relationship-building skills. While none of this replaces the importance of their technical training or their motivation to use their talent in their projects, it does show how we can dramatically enhance the quality of their work which will continue to make our world better.\n",
"role": "user"
},
{
"content": "Purpose: Little research has explored the importance of interpersonal skills, and more specifically, emotional and social intelligence (ESI) competencies for an engineer's effectiveness or engagement. Furthermore, to the knowledge, no studies have explored the explanatory power of ESI over and above general mental ability and personality for engineers. The paper aims to discuss these issues.\nDesign/methodology/approach: In this study the authors gathered multi-source data for 40 engineers in a multi-national manufacturing company.\nFindings: The authors found that ESI as observed by their peers significantly predicted engineer effectiveness (DR2=0.313), while general mental ability (g) and personality did not. In the same study, an engineer's engagement in their work was significantly predicted by the degree of shared vision within their teams, while g, personality and ESI did not predict engagement.\nResearch limitations/implications: The authors explore the implications of the findings for corporate training and development, undergraduate education, and graduate education of engineers.\nOriginality/value: The authors draw on 30 years of longitudinal studies showing ESI and quality of relationships can be significantly improved with the appropriate pedagogy emphasizing the building of one's vision, developmental approaches to ESI, developing a shared vision with others, and inspirational coaching.\n",
"role": "assistant"
}
] |
The influence on corporate entrepreneurship of technological variables
|
[
"Spain",
"Technological distinctive competencies",
"Top management support",
"Organizational learning",
"Corporate entrepreneurship",
"Organizational performance"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: The 1990s might well be considered the beginning of the era called the new or e-economy because of the appearance of the internet in the economy. In today's globally competitive environment, the internet has led to tremendous expansion in technology for business and rapid growth in investments (Skerlavaj and Dimovski, 2006).Assessing the value of technology has never been easy. The application of technology throughout the company (Andreu and Ciborra, 1996; Giarratana and Torrisi, 2010; Leonard-Barton, 1992) has increased the importance of technological distinctive competencies (TDCs), which permit the exploitation of the technological opportunities for the development of corporate entrepreneurship, reduction of costs, and generation of organizational competitive advantage (Alvarez and Barney, 2007; Autio et al., 2000; Real et al., 2006; Woolley, 2010).In this research, TDCs represent \"the organization's expertise in mobilizing various scientific and technical resources through a series of routines and procedures which allow new products and production processes to be developed and designed\" (Real et al., 2006, p. 508).This study contributes to prior research by linking TDCs to other strategic constructs, such as the necessity of top management support (TMS) for technology to obtain these competencies and the learning processes needed throughout the organization, that analyze globally and empirically their repercussions for corporate entrepreneurship in order to achieve a higher value firm. The study thus first seeks to analyze empirically the influence of TMS on TDCs (Byrd and Davidson, 2003; Haro-Dominguez et al., 2010). Second, we seek to analyze how these TDCs and TMS influence organizational learning (OL) (Alvarez and Barney, 2007; Autio et al., 2000; Leonard-Barton, 1992; Shane and Venkataraman, 2000; Teece, 1986; Woolley, 2010). Third, we analyze the direct relationships between both TDCs and OL on corporate entrepreneurship (Alvarez and Barney, 2007; Leonard-Barton, 1992; Woolley, 2010). Organizations that engage in entrepreneurial activities achieve higher levels of growth and profitability than organizations that do not, thus also achieving higher organizational performance (Antoncic and Hisrich, 2001). Corporate entrepreneurship involves \"extending the firm's domain of competency and corresponding opportunity set through internally generated new resource combinations\" (Covin and Slevin, 1991, p. 1). Thus, corporate entrepreneurship refers to \"the process by which firms notice opportunities and act to creatively organize transactions between factors of production so as to create surplus value\" (Jones and Butler, 1992, p. 735).All of these technological variables demand strong and committed TMS to guide the initiative and develop a working environment that supports technology (Ghosh et al., 2001). TMS \"reflects, in many ways, the importance that top management executives place on technology\" (Byrd and Davidson, 2003, p. 246). For Leonard-Barton and Deschamps (1988, p. 1254), TMS is a \"perceived powerful source\".TDCs influence corporate entrepreneurship, improving organizational performance. They also impact corporate entrepreneurship directly because they permit entrepreneurs to earn their status by demonstrating excellence in technological and professional skills and knowledge (Leonard-Barton, 1992). TMS has been closely linked to TDCs (Byrd and Davidson, 2003; Ghosh et al., 2001; Leonard-Barton and Deschamps, 1988).Although corporate entrepreneurship is directly influenced by TDCs, it may also be influenced indirectly through OL (Benitez-Amado et al., 2010; Chenhall, 2005; Gonzalez-Alvarez and Nieto-Antolin, 2005; Leonard-Barton, 1992; Omerzel and Antoncic, 2008; Rerup, 2005), because OL disseminates the knowledge acquired and uses it to encourage entrepreneurial behaviour (Garcia-Morales et al., 2006; Senge et al., 1994). OL can be considered a process whereby members of an organization are stimulated to strive continually for new approaches and to acquire, as well as to share, knowledge that influences their interactions with their environments (Argyris, 1993; Nonaka and Takeuchi, 1995).In the section on hypotheses, we draw on prior research to develop a number of testable hypotheses concerning the influence of TMS on TDCs, the influence of TMS and TDCs on OL, the influence of TDCs and OL on corporate entrepreneurship, and the way corporate entrepreneurship improves organizational performance. The section on hypotheses constitutes the theoretical foundation for the paper. The following sections present the research methodology and discuss the results. The final section makes some concluding observations and points out some of the study's limitations and lines for future research.\n2 Hypotheses: 2.1 The influence of TMS on TDCs and OL\n3 Methodology: 3.1 Sample and procedure\n4 Results: In this section, we present the main research results. First, Table III shows the means and standard deviations as well as the inter-factor correlation matrix for the study variables. There are significant and positive correlations among TMS, TDCs, OL, corporate entrepreneurship and organizational performance.Second, a structural equations model was performed to estimate direct and indirect effects using LISREL with the correlation matrix as input. This type of analysis has the advantage of correcting for unreliability of measures and also gives information on the direct and indirect paths between multiple constructs after controlling for potentially confounding variables. Figure 2 shows the standardized structural coefficients. The relative importance of the variables is reflected by the magnitude of the coefficients.Concerning the quality of the measurement model for the sample, the constructs display satisfactory levels of reliability, as indicated by composite reliabilities ranging from 0.84 to 0.92 and shared variance coefficients ranging from 0.57 to 0.87 (Table IV). Convergent validity can be judged by observing both the significance of the factor loadings and the shared variance. The amount of variance shared or captured by a construct should be greater than the amount of measurement error (shared variance>0.50). All of the multi-item constructs meet this criterion, each loading (l) being significantly related to its underlying factor (t-values>15.55) in support of convergent validity. To assess discriminant validity, we performed a series of kh2 difference tests on the factor correlations among all constructs (Anderson and Gerbing, 1988). We did this for each pair of latent variables by constraining the estimated correlation parameter between them to 1.0 and then performing a kh2 difference test on the values obtained for the constrained and unconstrained models (Anderson and Gerbing, 1988). The resulting significant differences in kh2 indicate that the constructs are not perfectly correlated and that discriminant validity is achieved.The overall fit measures, multiple squared correlation coefficients of the variables (R2s), and signs and significance levels of the path coefficients all indicate that the model fits the data well (kh2242=574.86; p>0.001; khratio2=2.56; NFI=0.99; NNFI=0.99; GFI=0.99; CFI=0.99; IFI=0.99; PGFI=0.80). The hypothesized model was a significantly better fit than the null model (kh2532=11,984.91; p>0.001; Dkh292=11,410.05; p>0.001). All modification indices for the beta pathways between major variables were small, suggesting that additional paths would not significantly improve the fit. The residuals of the covariances were also small and centred around zero.If we examine the standardized parameter estimates (Table V), the findings show that TMS (g11=0.82; p<0.001) is related to and affects TDCs, as predicted in H1. The model explains TDCs well (R2=0.67). TMS affects the development of TDCs because human, conceptual, and technological competency are thought to be supported by managerial effectiveness (Benitez-Amado et al., 2010). TMS also affects OL directly (b21=0.42; p<0.001). Furthermore, we have shown an indirect effect (0.33; p<0.001) of TMS on OL by TDCs (0.82x0.40; see Bollen (1989) for calculation rules). The global influence of TMS on OL is thus 0.75 (p<0.001), supporting H2. TMS encourages development of OL processes, which enable generation of new technological abilities and knowledge (Garcia-Morales et al., 2006). TDCs affect OL (g21=0.42; p<0.001), supporting H3. TDCs also allow an organization to increase its capability to carry out actions that improve OL (Senge et al., 1994). Comparing the magnitudes of these effects indicates that the effect of TMS on OL is larger than the effect of TDCs on OL. Globally, OL is explained well by the model (R2=0.61).Corporate entrepreneurship is influenced by TDCs (b31=0.43; p<0.001) and OL (b32=0.47; p<0.001), supporting H4 and H5, respectively. Furthermore, we have shown an indirect effect (0.19; p<0.001) of TDCs on corporate entrepreneurship by OL (0.40x0.47). The global influence of TDCs on corporate entrepreneurship is thus 0.62 (p<0.001). Comparing the magnitudes of these effects indicates that the total effect of TDCs on corporate entrepreneurship is larger than the total effect of OL on corporate entrepreneurship. Globally, corporate entrepreneurship is explained well by the model (R2=0.71).TDCs and OL enhance the capacity and knowledge needed to discover and exploit entrepreneurial opportunities (Omerzel and Antoncic, 2008). Finally, we find a significant relationship of organizational performance with corporate entrepreneurship (b43=0.64; p<0.001), supporting H6. Corporate entrepreneurship can increase companies' profits (Andreu and Ciborra, 1996; Shane and Venkataraman, 2000). Globally, the model explains organizational performance well (R2=0.41). Other indirect effects are shown in Table V.In testing the theoretical framework, we fit several nested models, each incorporating different assumptions about parameters. Comparisons with reasonable alternative models are recommended as a means of showing that a hypothesized model is the best representation of the data. Comparison is an important part of assessing model fit (Bollen and Long, 1993). The summary statistics in Table VI indicate that model 1 is preferred to the others, supporting the inclusion of a model with these relationships among the analyzed constructs. For example, if we compare the theoretical model (model 1) to a model that does not consider the relationship between corporate entrepreneurship and organizational performance (model 6), we see that the latter has a worse root mean square error of approximation (>RMSEA=0.008), expected cross-validation index (>ECVI=0.30), Akaike information criterion (>AIC=58.92), consistent Akaike information criterion (>CAIC=58.92) and estimated non-centrality parameter (>NCP=58.92). Hence, the results show that corporate entrepreneurship affects organizational performance and that model 1 is preferred to model 6 (Dkh2=58.92; Ddf=1). The theoretical model is also preferable to the other models formulated (Table VI). Length restrictions prevent detailed discussion of each model and of other models (a full report is available from the authors). In sum, the proposed theoretical model represents (Figure 2) the preferred, i.e. the most acceptable and parsimonious, model.\n5 Conclusions and future research: 5.1 Discussion\n",
"role": "user"
},
{
"content": "Purpose: - The aim of this paper is to highlight the importance of different technological aspects of organizations on Spanish high-technology firms' performance.\nDesign/methodology/approach: - The relationships studied are confirmed empirically using a structural equation model to demonstrate our hypotheses. The sample was selected from the database \"Dun & Bradstreet Espana\" in the year 2005 and includes 201 Spanish firms. CEOs were our main informants.\nFindings: - The results obtained show that support from top managers will directly influence the organizational learning (OL) process and technological distinctive competencies (TDCs) (antecedents of corporate entrepreneurship) and that corporate entrepreneurship finally influences organizational performance.\nResearch limitations/implications: - The paper is exploratory in character, and its goal is to show whether interrelations exist between the variables. The main limitations are: the sectors chosen refer only to Spain; the analysis is cross-sectional in character; the study uses a single method and self-reports (CEOs).\nPractical implications: - To obtain perfect adaptation of the firm to its environment, it is crucial that managers develop corporate entrepreneurship to improve high-technology sector firms' performance. The paper shows the important role of the top manager's support in developing TDCs and OL. Success in such issues is of vital importance to corporate entrepreneurship in the firm.\nOriginality/value: - The paper seeks to stimulate new lines of research on one variable (TDCs) and to relate it to other constructs, producing new relationships and observing their repercussions for the firm.\n",
"role": "assistant"
}
] |
Blood transfusion policy in Dutch hospitals
|
[
"Blood transfusions",
"Quality assurance",
"Hospitals",
"The Netherlands"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In The Netherlands, Consensus Development Conferences (CDC) are organised by the National Organisation for Quality Assurance in Hospitals (CBO). CBO is an independent organisation founded in 1979 by the Medical Specialists' Organisation and the Medical Association of Hospital Directors. Its purpose is to stimulate and guide quality assurance activities in hospitals.One of the elements of quality assurance is the establishment of criteria for good health-care delivery. During the implementation of a quality assurance programme in hospitals, CBO became aware of the physicians' need for national guidelines concerning a number of medical topics. This led CBO to start a consensus development programme (CDC) in 1982. Up to 1987, 22 CDCs have been held. Details of the CDC procedure employed, the choice of topics, and the differences between our approach and those of other countries have been described elsewhere (Nederlands Tijdschrift voor Geneeskunde, 1983; Casparie and van Everdingen, 1985). One of the distinctive characteristics of CDC in The Netherlands is that they have originated in the field of quality assurance rather than in the field of technology assessment. Besides assessment of guidelines and criteria for quality assurance during a CDC, the importance of the implementation of the conclusions in the daily practice of care must also be recognised. We therefore conducted an evaluation of our first CDC on blood transfusion policy (Casparie et al., 1987) to assess the effect of this CDC on the daily practice of care and on the quality assurance activities in Dutch hospitals. The results of an analysis of these studies were presented at the Third International Symposium on Quality Assurance in Health Care held in Paris in November 1986 (van Everdingen et al., 1986), and are discussed in this article.\nCDC on blood transfusion policy: The first CDC was held in 1982 in The Netherlands and addressed the topics of blood-ordering and transfusion policies in hospitals (Casparie et al., 1987). Transfusion therapy was considered a costly and frequently applied form of treatment with potential risks and complications.Many hospitals did not have well-defined guidelines for a number of aspects of their transfusion policy. These were the main reasons for the choice of this policy as a topic for consensus development.A working committee consisting of 12 experts prepared statements concerning good health-care delivery with respect to blood transfusion therapy. During the conference on 9 October 1982 the members of this committee gave short presentations and defended the statements before an audience. The conference was attended by 320 participants, mainly internists, heads of bloodbanks, clinical chemists, surgeons and anaesthesiologists. The final consensus text was formulated by the working committee a few weeks after the meeting. This text was composed of the statements agreed on in the public discussion, together with explanatory remarks. The statements were published in the national medical journal (Nederlands Tijdschrift voor Geneeskunde, 1983).Elements of the blood transfusion policy on which explicit recommendations were made in the consensus text included:* identification of the blood samples;* double testing of the cross-match technique for missed compatibilities;* type and screen strategy and registration of irregular antibodies against red blood cells;* use of standardised ordering lists for the use of blood in surgery;* proportion of whole blood versus red-cell concentrates;* quality aspects of red-cell concentrates;* cross-match/transfusion ratio;* indications for use of blood and blood products;* administration of blood to the patient;* blood transfusion reactions; and* transfusion committee in hospitals.\nEvaluation of CDC: methodology: The evaluation took the form of a survey combined with interviews and an analysis of peer review studies. Three years after the conference a questionnaire was filled in by the heads of the Dutch bloodbanks (number =22). The annual reports of the bloodbanks between 1981 and 1986 supplied quantitative data. A structured interview comprising 56 questions was held with a representative sample of heads of hospital haematological laboratories (57 out of 183). After the conference a quality assurance study on blood-transfusion policy was performed in 17 of the 183 Dutch hospitals.\nEvaluation of CDC: results: Questionnaire sent to 22 heads of bloodbanks\nDiscussion: Assessment of guidelines and criteria for good health care is not sufficient for adequate improvement of the practice of care. Implementation of these guidelines in daily practice, and thus effective improvement of care, must be the ultimate goal. Systematic evaluations of the effect of CDCs on the practice of medical care have not been published yet, but some preliminary results from various countries that conduct CDCs were recently presented at the Third Annual Meeting of the International Society of Technology Assessment in Health Care in Rotterdam (Calltorp, 1987; Hill, 1987; Kanouse and Jacobi, 1987; Klazinga et al., 1987; Lomas et al., 1987). That a CDC can have an impact is suggested by a recent study done in the USA, where the trial of labour after a previous caesarean delivery, as advocated in the NIH consensus statement, increased almost fourfold from 2.1 per cent in 1979 to 8.0 per cent in 1984.For our CDC method we have chosen a procedure with two group processes. The first of these group processes is embodied in meetings by a committee of experts who will convene six to ten times until agreement is reached. The second process will occur during the conference and take the form of open discussion between the experts and the physicians and other health-care workers involved in that area of care, all of whom can influence the final consensus text. The experts are chosen in consultation with their specialist organisation, and we assume that their prestige will increase the credibility of the consensus development process.The announcement of a conference is sent to all physicians and health-care workers involved, and the conferences are attended by a substantial proportion of them (100-1,000 participants. The final consensus text is published in the national medical journal and is given large-scale distribution by the CBO. We think that our procedure ensures that the final consensus statements have a solid basis and will be accepted by the medical profession. These expectations were confirmed by the finding that the text of the CDC on blood-transfusion therapy was widely known and had been discussed in almost all hospitals in The Netherlands. Furthermore, guidelines and criteria had been adopted by the transfusion committees where necessary. In hospitals where transfusion policy was chosen as a topic for audit, the criteria applied were in agreement with the consensus statements. Thus, these guidelines can be considered to be of practical value.In only four of the 15 completed quality assurance studies did the current practice of care satisfy the criteria. Five out of seven re-evaluation studies showed improvement.Although other factors such as Aids and budgeting of hospitals undoubtedly also influenced blood-transfusion policy in The Netherlands, our findings suggest that the CDC on transfusion policy had a substantial impact on the daily practice of care.\n",
"role": "user"
},
{
"content": "Purpose: - The paper aims to investigate the quality in health care with regard to the blood transfusion policy in Dutch hospitals.\nDesign/methodology/approach: - The value of an approach to good standards of health care practice, based on Consensus Development Conferences is analysed in the paper in the context of blood transfusion policies.\nFindings: - The paper finds that the Consensus Development Conference on transfusion policy had a substantial impact on daily practice of care in Dutch hospitals.\nOriginality/value: - The paper provides useful information on maintenance of good health care standard with regard to blood transfusions in Dutch hospitals.\n",
"role": "assistant"
}
] |
Leading change through your creative class
|
[
"Organizational design",
"Talent management",
"Creative class",
"Organizational transformation",
"HR innovation",
"Unconference"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: From a pure academic point of view, the paper you are going to read is not valid; it tells how we are in the middle of distorting, with no proven success so far, a highly disputed theory to implement it in a challenging, non-conventional organisation in an industry which cannot be easily transposed to any other[1]. And, to make the fraud complete, to keep it as secret as possible.\nTalent, technology and tolerance: Florida considers that the attractive environment for the creative class is a mix between three elements: talent, technology and tolerance, the combination of which enables cities to attract the creative class. At Mazars, the three elements are present to various degrees across the organisation, but in major flux. Indeed, all three represent high-stakes issues that will impact our long-term success.\nEffectively empowering the creative class: Discern your creative class\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to share how a professional service firm transposed Richard Florida's \"Creative Class\" concept from the urban environment into a corporate one for the purpose of organisational change. The validity of Florida's concept is not here reviewed; rather, the paper is a case study on how his theory - that talent, technology and tolerance compose the high-value triptych driving a city's growth and attractiveness - can be appropriated by HR to trigger profound changes in corporate governance and culture.\nDesign/methodology/approach: The paper provides a return on experience of a project that was implemented at Mazars, an international mid-cap of 20,000 people in 86 countries, over the course of one year. Approximately, 50 individuals participated in the initiative, lead by the firm's HR leadership team, which used an under the radar approach based on a revamped version of Owen's Open Space Technology. From an academic perspective, the approach is inspired by Argyris' action science, and more specifically a derivation of the \"double loop learning\" where the initial intent of the research might be modified by intermediary findings.\nFindings: The paper offers a model for identifying the members of the \"creative class\" in a corporate environment and a tested approach for integrating the \"creative class\" into the exclusive and institutional exercise of setting strategy. The cumulative effect of this \"unofficial\" operation is the creation of unique thought leadership and projects, some of which have now been officially adopted in the four-year strategic plan and institutionalized in the new governance system, results difficult to achieve through conventional approaches.\nResearch limitations/implications: The case study, which is still in progress, has been implemented in a non-conventional organisation in a very specific industry.\nOriginality/value: To the authors' knowledge, this is the first application of Richard Florida's urban renewal theory in the corporate environment. This is an example of innovative HR management responding quickly and effectively to the digital, disrupted business landscape. It is designed in the modern managerial spirit of test-and-learn, structured as an agile initiative in an open-source world. It provides a prototype to be replicated and tested in other environments.\n",
"role": "assistant"
}
] |
The anatomy of business failure: A qualitative account of its implications for future business success
|
[
"Business failure",
"Entrepreneurship",
"Consequences",
"Interpretative phenomenological analysis",
"Learning from failure",
"M21",
"L26"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Business failure (BF) is a constant in today's business world, being considered an essential and significant part of new business ventures (Ucbasaran et al., 2013; Walsh and Cunningham, 2016). From the extant literature on the topic of the costs bared by the entrepreneurs, it is undeniable that BF is essentially a learning process (Cope, 2011).\n2. Research synthesis: 2.1 Defining BF\n3. Methodology: The method of qualitative research employed in this study consists of a set of case studies, using the methodology design advanced by Yin (2009), based on a collection of personal accounts from individuals that meet pre-defined criteria through face-to-face interviews, using the Interpretative Phenomenological Analysis (IPA) (Smith and Osborn, 2007). The IPA approach has been used before in the research on \"BF Consequences\" (Cope, 2011) and enables an introspective view of experiencing failure. Cope's (2011, p. 608) use of interpretative phenomenological research is justified with \"the strength of a qualitative research design such as this lies in its capacity to provide situated insights, rich details and thick descriptions. Richness is provided by paying close attention to both context and process [...].\" This methodology allows the researcher to empirically achieve a better understanding of the emotional consequences of failure and its relation with the learning process.\n4. Results: 4.1 How the individual progresses and eventually develops new ventures\n5. Conclusions: 5.1 Main contributions of the study\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the aftermath of business failure (BF) by addressing: how the individual progressed and developed new ventures, how individuals changed business behaviors and practices in light of a failure, and what was the effect of previous failure on the individual's decisions to embark on subsequent ventures.\nDesign/methodology/approach: The authors resort to qualitative methods to understand the aftermath of BF from a retrospective point of a successful entrepreneur. Specifically, the authors undertook semi-structured interviews to six entrepreneurs, three from the north of Europe and three from the south and use interpretative phenomenological analysis.\nFindings: The authors found that previous failure impacted individuals strongly, being shaped by the individual's experience and age, and their perception of blame for the failure. An array of moderator costs was identified, ranging from antecedents to institutions that were present in the individual's lives. The outcomes are directly relatable to the failed experience by the individual. The authors also found that the failure had a significant effect on the individual's career path.\nOriginality/value: While predicting the failure of healthy firms or the discovery of the main determinants that lead to such an event have received increasingly more attention in the last two decades, the focus on the consequences of BF is still lagging behind. The present study fills this gap by analyzing the aftermath of BF.\n",
"role": "assistant"
}
] |
The extension programme and food safety education in the USA
|
[
"Food safety",
"Grants",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nBackground: Most colleges and universities engage in research and teaching. In the United States, there are more than 100 land-grant colleges and universities that also have a mission of extension, which is the term used for the way that land-grant institutions \"extend\" their resources to solve public needs with college or university resources through non-formal, non-credit programs. Extension programs are largely conducted through thousands of county and regional extension offices located throughout the USA and its territories.The land grant universities and the local extension offices are supported by the United States Department of Agriculture's Cooperative State Research Education and Extension System (CSREES), the federal partner in the Cooperative Extension System. CSREES plays a key role in the land-grant extension mission by distributing Congressionally appropriated funding to supplement state and county funds.The US Congress created the extension system nearly a century ago to address exclusively rural, agricultural issues across America (CSREES, 2006). At that time, more than 50 percent of the USA population lived in rural areas, and 30 percent of the workforce was engaged in farming. Extension's engagement with rural America helped make possible the American agricultural revolution, which dramatically increased farm productivity.Fewer than 2 percent of Americans farm for a living today, and only 10 percent of Americans now live in rural areas. Yet, the extension service still plays an important role in American life - rural, urban, and suburban. With an extension office in or near most of the nation's approximately 3,000 counties, the Cooperative Extension System remains an important player in American life by addressing a wide range of human, plant, and animal needs in both urban and rural areas. Extension works in six major areas:1. family and consumer sciences;2. 4-H Youth development;3. agriculture;4. leadership development;5. natural resources and environmental management; and6. community and economic development.Family and Consumer Sciences programs include food safety education, nutrition, and food selection, preparation, and storage.Since 2000, CSREES has awarded integrated research, education, and extension grants that address a broad spectrum of food safety concerns through the National Integrated Food Safety Initiative (NIFSI), a competitive food safety grant program. NIFSI grant projects are uniquely positioned to address major changes in the food safety system because of they are focused on integration of the resources and expertise of government, industry, universities, public and private organizations, and consumers to solve complex food safety problems. NIFSI awards $15 million annually in competitive food safety grants.The overall purpose of NIFSI is to support competitive food safety grants that use an integrated approach to solving complex food safety problems. By definition, integrated grants include research, education, and extension activities, each of which complement each either and greatly increase the likelihood for success of food safety education efforts. The crucial key here is that integrated, competitive grants allow USDA and CSREES to fund food safety education efforts that move from theory, to data gathering, to data analysis, to testing, to program development, to implementation and intervention, and finally, to evaluation. This comprehensive approach provides the basis for an overall integrated food safety program. Applicants are strongly encouraged to assemble project teams that include applied research, education, and extension experts who will work collaboratively to develop each proposal submitted to the NIFSI program.\nOutputs: Since its inception in Fiscal Year 2000, NIFSI has awarded more than 180 competitive, integrated, food safety grants. Grants are awarded for up to 4 years, with the average grant funded at $550,000. On-going integrated activities are widely varied and include:* conducting workshops, seminars, conferences, national and international meetings;* developing interactive websites that provide answers to commonly asked consumer food safety questions;* curriculum and course development for education and training of a variety of food safety audiences;* publications in refereed journal articles;* citations in professional journals;* training of migrant farm workers across the USA using modified GAPS (Good Agricultural Practices) guidelines;* HACCP (Hazard Analysis and Critical Control Points) training and education for meat and poultry processors;* HACCP model development for fruit and vegetable growers; and* facilitating partnerships among federal, university, and industry partners.Target audiences are those who have participated in, or been impacted by NIFSI programs. Those audiences include, but are not limited to:* farm workers and growers;* producers and processors;* packers, shippers, and distributors;* wholesalers and retailers;* foodservice workers at restaurants, supermarkets, hospitals, schools, etc.;* volunteer food handlers at community kitchens, food pantries, health fairs, etc.;* consumers (youths, teens, 4-H clubs, adults, older Americans, etc.);* high-risk audiences (pregnant or nursing mothers, the very young, the very old, the chronically ill, and those who are immunocompromised);* multipliers (teachers, trainers, and other educators); and* undergraduate and graduate students in the food and agricultural sciences.\nOutcomes: Federal data from the Centers for Disease Control and Prevention have indicated that, because of a high level of commitment by multiple collaborators, there has been a 20 percent decrease in foodborne illness incidence over the past 7-8 years. This decrease is due to the collaborative efforts of federal, state, and local governments, professional organizations, industry and trade groups, and university researchers and educators, among others. No single agency or organization can be identified as the primary determinant for the decrease in illness incidence. Nevertheless, collectively, all collaborating partners and stakeholders have played a crucial role in influencing the results. NIFSI has contributed to this overall decrease in foodborne illness incidence by using a unique and innovative approach not currently employed by any other collaborating partner - an integrated approach that brings together the resources of the government, universities, and the food industry to solve complex food safety problems.The National Integrated Food Safety Initiative (NIFSI) has solicited grant applications that focus on building and enhancing national support and coordination of integrated food safety programs and resources. Funded grants support the development of databases and interactive software that aid in decision-making and problem solving among food safety researchers, educators, consumers, and others. International, national, and regional conferences, multi-user distance education programs, satellite video-teleconferences, and other communication transfer technologies have also been funded. Other NIFSI grants have sought to improve food safety education and outreach to high-risk, under-served, or hard-to-reach audiences - those who are most at risk for developing foodborne illnesses.NIFSI has awarded more than $700,000 in grant funds to university collaborators working in conjunction with the national Partnership for Food Safety Education and its Fight Bac campaign (Bac is a green, cartoon character - a germ or bacteria - who teaches users how to recognize contaminated food and food surfaces). The University of Georgia developed a Fight Bac interactive computer module for elementary school students with NIFSI funding.NIFSI has awarded more than $1.5 million for the development of interactive databases and interactive CD-ROMS that target child care providers, foodservice workers, teachers, health care workers, and consumers. For example, at Iowa State University, investigators have developed a virtual food safety reference desk. Consumers, teachers, health care workers and others receive cutting-edge, science-based responses to their food safety questions from juried panels of food safety experts. The database receives an average of 2000 hits each month. The volume of questions has been so great that investigators have developed web-based FAQ pages and subject matter Q&A publications based on commonly asked questions.According to the Centers for Disease Control and Prevention, the recent 20 percent decrease in the incidence of foodborne illness in the USA is also attributable, in part, to the adoption of HACCP programs in meat and poultry plants throughout the USA. In the early 1990s, USDA mandated HACCP programs for meat and poultry processors of all sizes. This posed a significant problem for small and very small meat and poultry plant owners, many of whom had limited access to adequate HACCP training and implementation. NIFSI worked cooperatively with federal partners in the Food Safety and Inspection Service to help build and establish training programs for these smaller facilities, primarily through joint training partnerships with university experts. NIFSI has provided more than $22 million in grant funds to support HACCP program development and implementation across the USA.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to explain the work and importance of the \"extension\" programme in the USA.\nDesign/methodology/approach: - An historical review of the formation, development, present day workings and importance of the extension programme.\nFindings: - \"Extension\" works in six major areas, including family, consumer science and agriculture. Extension grants that address food safety are issued via the National Integrated Food Safety Initiative (NIFSI). Target audiences for the programmes include various groups of consumers (pregnant/nursing mothers, infants, teens, etc.), as well as multipliers or gatekeepers who disseminate knowledge to groups of consumers.\nPractical implications: - Since the Year 2000, NIFSI have awarded more than 180 competitive integrated food safety grants. Federal data indicate 20 percent reduction in foodborne disease attributed to collaborative educational efforts including the development of HACCP.\nOriginality/value: - This paper is one of the few to describe the work of the extension programme with specific reference to food safety.\n",
"role": "assistant"
}
] |
Critical times in business relationships
|
[
"Business policy",
"Professional services"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Concepts like \"critical incidents\" (for an overview, see Edvardsson and Roos (2001), Roos (2002) and Gremler (2004), \"critical events\" (Halinen, 1994; Halinen et al., 1999; Havila and Salmi, 2000), \"precipitating events\" (Tahtinen, 2001) and negative incidents (Holmlund and Strandvik (2005) and Holmlund-Rytkonen and Strandvik (2005)) have been used to capture single or combinations of interactions that may have an effect on perceptions and behaviour in the relationship between a seller and a buyer. Recent research shows, however, that changes in business relationships are often the result of a longer and more complex process. In this paper, we introduce the concept \"critical time\" as a period of stress and raised sensitivity in a relationship that may change it. The traditional incident-based view thus becomes an embedded part of critical time. Further more, our results from an empirical study of two professional services indicate that the proposed conceptual framework can be used to identify and understand critical times in business relationships.We define \"critical time\" as a period of time with increased sensitivity in a business relationship that may change the actors' attitude and/or behaviour in the relationship. It starts with either actor's experience that something in the relationship does not meet their expectations and ends when this issue is settled. Critical times are built up by three elements embedded in a temporal and situational context: initial state of the relationship, the flow of critical time and outcome state of the relationship. By introducing the critical time perspective, time can be included as well as the context both within the relationship, the external business environment and the internal organizational environment. Our paper represents a reframing of the traditional critical incident (CI) perspective by widening the scope to include time and context-related dimensions. We argue that focusing on isolated incidents is not enough if we want to identify and understand dynamics in business relationships.The traditional CI approach has been widely used to study service episodes embedded in consumer relationships (Gremler, 2004). A number of developments of the original ideas by Flanagan (1954) have been presented and argued for in the services literature. Most studies focus on the CI as such or on the cause, course and effect of isolated incidents (Edvardsson, 1988), and in only some studies there has been an explicit customer relationship focus and a contextual analysis. Edvardsson and Strandvik (2000) reported that incidents that traditionally would be defined as critical (by both the respondents and the scholars) are not critical for the customer relationship in the sense that they would result in a permanent change of the relationship. A single incident may thus not be a fruitful unit of study. A period in time might be more appropriate as a unit of analysis in order to understand how relationships develop and change.Holmlund and Strandvik (2005) developed a model for diagnosing and managing negative incidents in business relationships suggesting that recognizing what the customers experience as exceptional and disappointing can be considered essential in business relationship management. They introduce the term negative incidents arguing that:Negative incidents represent situations where companies' attention level and sensitivity are raised, and thus they may initiate actions that affect how a relationship develops. The new term covers negative and significant situations more broadly than previously in critical-incident and critical event studies and ranges from various interaction failures to different kinds changes (p. 227).In their diagnostic model, they similarly to Edvardsson (1988) distinguish between the context, which is the nature of the relationship where the incident occurs, the content that is the nature of the incident, the course which refer to actions caused by the incident and consequences that is effect on the relationship. Holmlund-Rytkonen and Strandvik (2005) further suggest aggregating all negative incidents to a stress measure and introduce the concept \"hot spots\" to denote areas in the company's business with an exceptional occurrence of negative incidents. They thus emphasize the importance of both temporal and contextual factors.Recent research reviewed in the next section concerning business relationships (Tahtinen, 2001; Halinen and Tahtinen, 2002; Tahtinen, 2002; Flint et al., 2002; Tikkanen and Alajoutsijarvi, 2002; Dubois et al., 2003) confirm this view.Our research focus is to explore the usefulness of the critical time conceptualization in order to understand business relationship dynamics. We study business-to-business (B-to-B) relationships concerning knowledge intensive professional services. The aim is to develop a theoretical framework for and an initial understanding of critical times in such business relationships. Our focus is on the factors driving the flow of critical times. In the following sections, we first review previous research; second, present our theoretical framework; and in the third section, we present the findings from two explorative empirical studies concerning companies' relationships with information technology (IT) consultants and advertising agencies. In the last section, we conclude by discussing the usefulness of our framework for studying contents and drivers of critical times in business relationships. We also suggest directions for future research.Previous research\nTheoretical framework and research approach: We argue that the focus on incidents does not seem to be sufficient to explain relationship dynamics, that is, the development from a current (which we will label: initial) state of the relationship to a new and changed state of the relationship when unexpected and unwanted events happen. Further more, we recognize that the development within a relationship or in the external business environment may develop gradually and result in changes in the relationship and also in relationship ending (Halinen and Tahtinen, 2002). Therefore, we argue that it is more relevant to use the concept critical time to understand the contents and drivers of relationship change and to depict the effects on the relationship in the future.Critical time starts when either the buyer or the seller perceives new challenges or emerging problems in the relationship that diverges from what has been expected. A critical time ends when the issue has been dealt with in a way that results in a stabilized relationship which may be strengthened, sustained, faded or a dissolved relationship. Critical time thus contains three elements:1. an initial state of the relationship;2. the flow of critical time; and3. an outcome state of the relationship.The context is embedded in the initial state of the relationship and will affect the flow of the critical time as well as the outcome state of the critical time.How long is a critical time? Tahtinen (2002) studied a long and winding ending process. We suggest that critical times may also be short and straight-forward resembling a dramatic incident. Holmlund (2004, 2001) discusses different observation units of interaction in a business relationship. Traditionally, the episode level and the relationship level have been used. She introduces additional levels, of which, the \"sequence level\" has particular relevance in our case. Holmlund argues that different interaction levels, action, episode, sequence and relationship represent a hierarchical system of perception units that are simultaneously present. The customer (and the supplier) may interpret a particular interaction simultaneously in different ways. Whether a disappointing action leads to an interpretation that the whole episode was disappointing, or even influences the interpretation of the next level, the sequence level, or the whole relationship is a multifaceted question. A sequence is defined as \"a time period, a product, a campaign or a project, or a combination of these\" within a relationship (Holmlund, 2004, p. 282). Holmlund's sequence concept is thus closely related to our \"critical time\" concept in the sense that critical time may be seen as periods of unrest between stable sequences. An interesting point in Holmlund's model is that she points to dialectical change in the relationship. The perception of a certain action is not only a CI as such but also may influence perceptions of other actions and episodes and even change the overall perception of the relationship, thus function as a filter how the next (similar) action is interpreted. Through her model she proposes that the perception of relationship quality and (changes in the perception) is the result of a complex process.How do critical times develop and why? What are the drivers of the flow of critical times? The flow of critical time refers to the responses from relationship partners as a consequence of triggers. The flow of responses includes interactions, communication, information, changed behaviour, stress, decisions and various emotional responses. The flow of responses initiated by one or a combination of triggers causing critical times which after a shorter or longer period will go back to normal, similar to a storm being over. Critical times turn to normal times as perceived by the actors involved. We have reached the outcome state of the relationship in our framework. The outcome state could be a strengthened, sustained, faded or dissolved relationship.We will in this paper explore the role of influencing factors in the flow of critical time in business relationships. These may include incidents or events such as quality problems in the services delivered, reorganization or strategic changes (including either the buyer or the seller or both), a new supplier offering a new and \"better\" service and ownership change but also factors that gradually develop over time such as lack of interest on the behalf of the service provider, communication problems and stress due to expectations not being fulfilled. We are not only interested in the triggering factors but also in the total flow of the critical time which includes reactions and responses from both the seller and the buyer as well as context-related factors within and outside the relationship, which may initiate, form and direct the flow of critical times.In the CI literature, trigger is often used to describe relationship dynamics. Triggers are often used to refer to, e.g. \"starters\" of processes, \"incidents\" or \"events\" that give energy or direction to process that deviates from what is defined or perceived as normal or in more general terms, triggers are \"alarm clocks\" (Gardial et al., 1996) which call for response either in organizations (Schindehutte et al., 2000) or in customer relationships (Roos, 1999; Roos and Strandvik, 1997; Edvardsson and Strandvik, 2000, 2002). Olsen (1992) defined a trigger as critical steps in an episode of a customer relationship. Edvardsson et al. (2002) used triggers as the reason behind a change of the relationship into more conscious and sensitive approaches to all activities and interactions.To summarize: our research approach is focused on identifying and model the underlying factors driving the flow of critical times from the initial state to the outcome state of business relationships. A content analysis of narratives describing the flow of critical times using constant comparative analysis inspired by a grounded theory approach will be carried out to create the basis for our new model of the combination of factors giving energy and direction to the flow of critical times in business relationships. The factors describe the contents to be focused on when managing critical times and thus the future of business relationships.\nEmpirical study: Two exploratory empirical studies in business settings were carried out concerning companies' evaluation of their IT-consultants and advertising agencies, respectively. The IT-study and the advertising agency study were both carried out in Sweden. Data were collected in personal interviews with employees in the buying company, being responsible for the relationship with the service providers and the services bought. The perspective is thus the business customers' perspective.Data collection\nDiscussion: The aim of this paper was to explore the usefulness of the critical time conceptualisation when it comes to understanding business relationship dynamics. From the outset, we contrasted this view to research focusing on CIs and critical events. Edvardsson and Strandvik (2000) reported that incidents that traditionally would be defined as critical are not necessarily critical for the customer relationship in the sense that they would result in a new state of the relationship (strengthened, weakened or dissolved). We reasoned that a period in time when the relationship changes from one state to another might be more appropriate as a unit of analysis. Recent research concerning business relationships gave indications in this direction (Tahtinen, 2001; Halinen and Tahtinen, 2002; Flint et al., 2002; Tikkanen and Alajoutsijarvi, 2002; Dubois et al., 2003). Therefore, we propose that it is relevant to use the concept critical time to understand how relationships change.We have defined critical time as a period of time of stress and raised sensitivity in the business relationship. It starts when the customer experiences a divergence from expectations about something in the relationship and ends when the issue is settled in some way. In the empirical study, we identified examples of critical time by asking the respondents to describe events that have caused disturbance in the relationship. Their description of critical times includes implicitly the time and context factors emphasized in the beginning of this paper. Context-related factors as perceived by the customers are thus embedded in the narratives of critical times. A single event or incident was often mentioned as a starting point but in the narratives it became clear that this usually was connected to a series of activities and responses until the issue was settled. This observation and result suggest the necessity of paying attention to time in order to understand customer relationship dynamics.This paper has two research contributions. First, it contributes with a new conceptualization; and the critical time framework. A critical time may change the actors' attitude and/or behaviour in the relationship but the outcome may also be no change. We suggest that critical time can be captured by three elements:1. initial state of the relationship;2. the flow of the critical time describing what gives energy and direction and the flow of responses which eventually result in an; and3. outcome state of the relationship.It would be plausible that the nature of the initial state of the relationship has an effect on the flow and the outcome. Three factors consistently involved in the flow of critical time stood, however, out, competence, communication and time. Each of these seemed to have two different modes similar to what has been proposed in service quality models, a process and an outcome mode. This means on one hand that customers have expectations about what they are buying but also on the other hand have expectations about the interaction within the relationship. This process aspect becomes accentuated when something unexpected and negative happens. It seems quite obvious that actions and responses in such situations are critical and the process may escalate or level off depending on many factors. From a managerial point of view, these process factors are thus significant capabilities.The second contribution of this paper is the critical C-model based on competence, communication and clock (time). Our empirical data indicates that each factor can be described in terms of process and outcome. This categorisation has been used in classic episode-based service quality models (Gronroos, 1984) and Holmlund's model of levels in business relationships where several service episodes affect the development of the relationship. Our model describes key factors in critical times and depicts graphically that there is the possibility for a negative or a positive relationship development resulting in either cooperation or in conflict. How this process evolves remains to be studied in further research including other empirical contexts and deepening the understanding of the role of the constellation of time, competence and communication during the flow of critical times.Managerial implications\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to develop a theoretical framework for and an initial understanding of critical times in business relationships.\nDesign/methodology/approach: - The authors define \"critical time\" as a period of time with increased sensitivity in a business relationship that may change the actors' attitude and/or behaviour in the relationship. The authors review previous research, present the theoretical framework and present the findings from two explorative empirical studies concerning companies' relationships with information technology consultants and advertising agencies. The focus is on the factors driving the flow of critical times.\nFindings: - This authors introduce the concept \"critical time\" as a period of stress and raised sensitivity in a relationship that may change it. Critical times are built up by three elements: initial state of the relationship, the flow of critical time and outcome state of the relationship. The authors contribute with the critical C-model: competence, communication, and clock.\nOriginality/value: - This is the first study focusing on critical times in business relationships. The authors study two categories of knowledge intensive professional services. A new theoretical framework for, and an initial understanding of, critical times in business relationships are suggested.\n",
"role": "assistant"
}
] |
Trajectory data mining: integrating semantics
|
[
"Decision making",
"Semantics",
"Trajectory",
"Data mining",
"Spatiotemporal"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Nowadays, we are living in a fast-paced, highly advanced, technological environment. With changes and advancements every day, organizations must be prepared and informed with the latest possible trends and applications in order to achieve competitive advantage. Recognizing the novel use of data mining techniques on moving object data can help an organization to flourish and prosper by providing valuable information otherwise unknown and unavailable to decision makers. In addition, due to the recent innovations in mobile, sensor, and global positioning system (GPS) technologies, capturing the evolving position of moving mobile objects has become both economically and technically feasible. Such advancements have opened the doors for many applications built on the knowledge of the specific movements of objects equipped with sensory devices (Spaccapietra et al., 2007). Moreover, as the data of moving objects constantly increases, a need has arisen to model, process, and mine this trajectory data, just like any other data before, in order to find and understand the patterns in these sets of data (Alvares et al., 2007b).\n2. Trajectory properties: In order to understand the topic of trajectory data mining, we must first understand what a trajectory is. Trajectories are described as the path, which an object traveling through space and time follows. Each point depicted on the path would represent a position in space as well as an instant in time (Alvares et al., 2007b). So trajectory data would normally be depicted as sample points in the form (tid; x; y; t), where tid is the identifier of the moving object, x and y are the object's coordinates in space, and t is the timestamp (Alvares et al., 2007a). Furthermore, trajectories are characterized by the speed of their movement, the headed direction of the trajectory, the area covered by the trajectory during its travel, the traveled distance, and the traveled time. Trajectory objects differ from any moving object, in that their movement is intended to satisfy a goal that requires the meaningful travel from one place to another in a finite amount of time (Brakatsoulas et al., 2004). In addition, trajectories can be described as \"the user defined record of the evolution of the position (perceived as a point) of an object that is moving in space during a given time interval in order to achieve a given goal.\"\n3. Trajectory data mining: Due to the wide availability of positioning technology, trajectory data has rapidly increased. As with all data, trajectory data also requires modeling, processing, and mining in order to find and understand patterns and relationships behind the data. Therefore, trajectory data mining has gained recognition in extracting useful information and knowledge from the data (Oda et al., 2009). Data mining is the process of applying algorithms in order to analyze sets of data, and extract useful and unknown patterns, relationships, and information (Adams, 2010). Trajectory data mining is the application of data mining techniques on trajectory data in order to extract patterns, which describe the behavior of the trajectories (Marketos, 2009).\n4. Semantic trajectory data mining: Existing work related to trajectory data mining has mainly focussed on the properties of the trajectories while disregarding the related semantics and background geographic information. However, if the geographic space where trajectories are located is to be considered, many more meaningful patterns which were not previously available can be extracted (Alvares et al., 2007b). Previous trajectory studies defined trajectory data as sample points without any semantic information, and trajectories as lines that a moving object follows through these points. These methods do not handle or analyze the geometric locations or semantic features associated to the trajectory, which are important to comprehending the trajectory data (Yan and Spaccapietra, 2009). The enrichment of trajectories with semantic geographical information could be very useful in simplifying the querying, analysis, and mining of the trajectory data (Alvares et al., 2007a). Furthermore, a semantic perception of the locations visited by a trajectory adds to the understanding of the moving object's behavior. These locations can further be clustered to extract significant locations of interest to the trajectory. Moreover, the semantic locations can be ranked by significance, according to several features such as the frequency and duration of the moving object's visits to that location (Cao et al., 2010). Such additions to the knowledge available provide a baseline for smarter applications. In the following sections we will discuss how trajectory data can be enhanced with semantic geographical information, as well as the processes applied for semantic enrichment.\n5. Research gap: The mainstream literature focussed on a limited scope of applications for trajectory data mining. The applications discussed most frequently were related to traffic management, animal migration, etc., without much reflection on other possible novelties, such as space and surgical applications. In the applications covered, the trajectory objects studied were mainly concrete and large physical objects, such as vehicles, people, and animals. Nevertheless, studies rarely considered the possibility of applying trajectory data mining to minuscule objects, such as atomic ions, or microscopic particles. However, due to the limitations in time and scope, this research will not focus on such applications, but future research can further explore these topics.\n6. Research methodology: In order to adequately investigate such a topic and answer the research question, an experimental quantitative approach is adopted. Throughout the paper, it has been shown that traditional trajectory data mining techniques do not handle or analyze the geometric locations or semantic features associated to the trajectory, and thus create a gap in the comprehension of the trajectory data (Yan and Spaccapietra, 2009). Subsequently, many more meaningful patterns which were not previously available through traditional techniques can be extracted only if the geographic space where trajectories are located is to be considered (Alvares et al., 2007b). Therefore, on a theoretical aspect, the application of data mining techniques on raw trajectories is not enough for enabling quality decisions. Furthermore, the semantic enrichment of trajectories and the provision of geographical details should allow for enhanced pattern extraction, as well as thorough and visible decisions on a more concrete basis. In the experiment, we aim to elucidate knowledge and put to practice by performing trajectory data mining and analyzing the results of pattern extraction and knowledge discovery.\n7. Results: The experiment was conducted on the set of raw trajectory data. The trajectory points were pre-processed and visualized using Weka's visualization module, with each of the ten cabs depicted in a different color. This visualization is represented in Figure 3. Next, a data mining technique was applied to the trajectory data. The simple expectation maximization (EM) clustering method was used in order to group the trajectories together based on similarity measures. EM goes through each trajectory instance and assigns it with a probability distribution to check the probability of that instance belonging to each of the clusters. The number of clusters to be created can be specified by the user, or determined by cross validation. The EM method clustered the trajectory data into ten different groups. The output and details of the clustering are shown in Figure 4. As we can see, the overall percentages of the cluster distributions are balanced. However, the low percentages in some clusters, such as clusters 1, 7, and 10, indicate a lack of similarity measures between the instances of these clusters and other instances. This may be due to some remaining outliers, despite the fact that the data has been pre-processed.\n8. Conclusion: In this paper we examined the innovative topic of trajectory data mining, which has gained much interest throughout recent years. In the first section of the paper, we discussed data mining in general in order to understand the data mining techniques and algorithms commonly used. After the data mining concepts were covered, the next section elaborated on moving object trajectories. The properties that define trajectories and the common types of trajectories were consequently explained. Subsequently, the third section examined the application of the discussed data mining techniques upon trajectory data. The extraction and mining of patterns from trajectory data was elucidated, and three applications of trajectory data mining were discussed in order to provide further clarification. Any new technology, if applied correctly can bring with it several potential benefits and innovations. Trajectory data mining is a remarkable field, which with its new applications and semantic enrichment techniques, seems to have opened many doors to innovation and a brighter future. After researching how data mining techniques can be applied to moving object data, we found that the typical data mining techniques commonly used, such as clustering, classification, and association techniques, can also be applied to trajectory data in a similar way. Data mining is used in general to extract unknown patterns from large sets of data, and applying it to trajectory data can help in discovering patterns and information of great importance to the decision maker.\n",
"role": "user"
},
{
"content": "Purpose: Trajectory is the path a moving object takes in space. To understand the trajectory movement patters, data mining is used. However, pattern analysis needs semantics to be understood. Therefore, the purpose of this paper is to enrich trajectories with semantic annotations, such as the name of the location where the trajectory has stopped, so that the paper is able to attain quality decisions.\nDesign/methodology/approach: An experiment was conducted to explain that the use of raw trajectories alone is not enough for the decision-making process and detailed pattern extraction.\nFindings: The findings of the paper indicates that some fundamental patterns and knowledge discovery is only obtainable by understanding the semantics underlying the position of each point.\nResearch limitations/implications: The unavailability of data are a limitation of the paper, which would limit its generalizability. Additionally, the lack of availability of tools for automatically adding semantics to clusters posed as a limitation of the paper.\nPractical implications: The paper encourages governments as well as businesses to analyze movement data using data mining techniques, in light of the surrounding semantics. This will allow, for example, solving traffic congestions, since by understanding the movement patterns, the traffic authority could make decisions in order to avoid such congestions. Moreover, it could also help tourism authorities, at national levels, to know tourist movement patterns and support these patterns with the required logistical support. Additionally, for businesses, mobile operators could dynamically enhance their services, voice and data, by knowing the semantically enriched patterns of movement.\nOriginality/value: The paper contributes to the already rare literature on trajectory mining, enhanced with semantics. Mainstream literature focusses on either trajectory mining or semantics, therefore the paper claims that the approach is novel and is needed as well. By integrating mining outcomes with semantic annotation, the paper contributes to the body of knowledge and introduces, with lab evidence, the new approach.\n",
"role": "assistant"
}
] |
Does university play significant role in shaping entrepreneurial intention? A cross-country comparative analysis
|
[
"Entrepreneurial intention",
"General educational support",
"Targeted cognitive support",
"University environment and support"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Looking at the roller-coaster ride various economies took in recent past, several stakeholders, including public policy makers and educational institutions, have reinforced the need to develop sound entrepreneurial culture to stabilize economies and create opportunities for growth. In parallel rather than seeking wage employment in highly dynamic and competitive corporate entities, increasing number of students now prefer to start their own venture based on a unique idea with minimal seed capital. Various researchers have found that new ventures established by university alumnae have significant economic multiplier effect in terms of job creation and income generation (Dietrich, 1999; Richert and Schiller, 1994; Carree and Thurik, 2005). Looking at need of society on one hand and growing aspirations of students as potential entrepreneurs on other, universities world-over are offering entrepreneurship courses (Postigo, 2002) and they actively provide mentoring support and assistance to promising students.\nReview of literature: Entrepreneurial intention and the TPB\nResearch methodology: Data reported in this paper were collected as part of a large study designed to test relationship between personal variables, contextual variables and situational forces that shape entrepreneurial intention. Convenience sampling method was employed, and survey was conducted among post-graduate management students in India, Malaysia and Singapore. Sample size was based on requirements for analysing the predictive model using hierarchical multiple regression, which was the statistical tool used for data analysis. There exist accepted rules of thumb to determine generalizable sample sizes when using multiple regression in a confirmatory fashion. For example, Hair et al. (1999) recommend a minimum five observations per independent variable - and desired number is 15-20 observations - so that the sample would be representative of population under study. According to this rule acceptable sample size for this study is 100 respondents. This calculation includes four independent variables namely entrepreneurial attitude, SN, PBC and university environment and support. Entrepreneurial intention is dependent variable (20x4=80+20=100). Since higher level of power for the study may be gained by increasing number of respondents, it was decided to have larger sample size than the minimum level for each country. Table I shows number of samples from the three countries.\nMeasurement: To measure first three independent variables of TPB, namely entrepreneurial attitude, SN and PBC, researcher has used the measure proposed by Linan and Chen (2006) in their Entrepreneurial Intention Questionnaire (EIQ) (see Table II for details on measurement of variables and reliability). This is a reliable and internationally applicable scale to measure TPB variables, for example in one of their studies which measured entrepreneurial intention using EIQ among Spanish and Taiwanese students, the composite reliability of 0.928, 0.892 and 0.919 for attitude towards entrepreneurship, SN and PBC are reported. In another study by Zampetakis et al. (2015), the same is reported to be highly reliable in context of Greek university students. Data for all four variables have been collected on a five-point Likert scale from \"strongly agree\" to \"strongly disagree\".\nReliability of measures: In order to assess reliability of measures in this study, item-to-total correlations and Cronbach's a were employed. As suggested by Nunnally (1978), criteria for retaining a scale item includes an item-to-total correlation of at least 0.35 and Cronbach's a of at least 0.70; however, Cronbach's a is allowed to go down to 0.6 since this is an exploratory research (Hair et al., 1999). SPSS produced separate internal consistency tests (i.e. reliability Cronbach's a test) for students from India, Singapore and Malaysia as well as for the whole data set. It is noted that samples from the three countries were homogenous and suitable for assessing reliability of the construct. Table II reports results of Cronbach's a for variables of TPB, that is entrepreneurial intention, entrepreneurial attitude, perceived SN and PBC. Cronbach's a for entrepreneurial intention for India, Singapore and Malaysian samples is 0.911, 0.931 and 0.956, respectively. With combined data set, overall Cronbach's a is 0.929 indicating satisfactory match with Nunnally's (1978) threshold.\nData analysis: To understand entrepreneurial intention and its relationship with university environment and support among students from India, Singapore and Malaysia, data are analysed keeping in view the hypotheses and major objectives set out for the study. First, a factor analysis is performed for university environment and support. Thereafter, structural equation modeling (SEM) is used to test the proposed model that postulates a relationship between entrepreneurial intention, attitude towards entrepreneurship, SN, PBC and university environment and support. Multivariate analysis of variance (MANOVA) is used to test the hypothesis to understand differences in perceptions of students from India, Singapore and Malaysia for university environment and support.\nFactor analysis for university environment and support: To understand underlying variables in the construct university environment and support, an exploratory factor analysis (EFA) is conducted. All measurement items for the same are examined for assumptions of factor analysis, such as normality, sufficient correlations and impact of influential observations (Hair et al., 1999). Correlation items of university environment and support items are examined to determine the appropriateness of EFA. Bartlett's test of sphericity is used to discover presence of correlations among underlying variables. It is significant at 0.000 for all items indicated that co-relationship matrix has significant correlations. Kaiser-Meyer-Olin measure of sampling adequacy showed satisfactory adequacy (0.955). Totally there are 18 items in the data. Two factors are extracted from these 18 items using the method of principal component analysis and rotation method of varimax, with criteria of eigenvalues greater than 1. These two factors explained 62.989 per cent of the variance.\nTesting entrepreneurial intention-university support model with SEM: It is considered that there are three considerations fundamental for interpretation of empirical analysis of a study of this nature: first, relevance of intentions in predicting and explaining behaviour, second, overall fit of models relevant to entrepreneurial activity, and third, relative strengths of significant entrepreneurial intention and activity antecedents. Considering this and the chief objective of study to derive a mathematical model to relate criterion or independent variable (students' entrepreneurial intention) to the predictors (ATB, SN, PBC and university environment and support), hypothesized model is tested using SEM technique. Formulation of the model is developed using AMOS 16.0 software package.\nDifferences in university environment and support: results of MANOVA and ANOVA: To determine whether students from India, Singapore and Malaysia differ significantly in their perceptions of university environment and support, a MANOVA was conducted. However, factor analysis conducted earlier has indicated that university environment and support is made-up of two factors, namely general educational support and targeted cognitive and non-cognitive support. Therefore, it was decided to conduct two separate MANOVA to understand the differences in perceptions related to these two factors.\nMajor findings and implications of the study: Considering socio-economic benefit of entrepreneurial activity among young students, academic institutes, government and other public-policy advocates are interested in understanding various factors that shape entrepreneurial intention so as to create conducive eco-system for new venture creation. For this, it is very necessary to understand psychological, situational and contextual factors that influence entrepreneurial intention (Tolentino et al., 2014). In this research, based upon theoretical framework of TPB (Ajzen, 1985), an attempt was made to understand influence of university environment and support on entrepreneurial intention of post-graduate management students of India, Singapore and Malaysia. The two factors that emerge from factor analysis of university environment and support are targeted cognitive and non-cognitive support and general educational support. Targeted cognitive and non-cognitive support accounts for 55 per cent of variance while general educational support accounts for almost 9 per cent of remaining variance. At educational support level, universities are expected to identify and nurture entrepreneurial trait among students and shape their competence, thus sowing seed and nurturing students' ambition to start a successful venture (Debackere and Veugelers, 2005). Along with this, a well-crafted entrepreneurship education curriculum can significantly raise students' enthusiasm and competence to become successful entrepreneur. At targeted cognitive and non-cognitive support level, as opined by Tijssen (2006), the emerging entrepreneurial universities are expected to play key role by providing support mechanisms like patents, technology transfer, and incubation among other necessary facilities to budding entrepreneurs. This will create a robust entrepreneurial eco-system and help students in converting their ideas into viable business models that may further expand into successful ventures, with due support mechanism from university and other institutes. In this way, university can become a nodal point of entrepreneurial activity among young target audience in their specific region and augment economic growth, prosperity and job creation. Thus broadly results of this study are in line with findings presented by Kraaijenbrink et al. (2010) that the university can be a strong support network and influencer in augmenting entrepreneurial activity among students.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to foster entrepreneurship among students and incubate more start-ups for economic prosperity, universities around the globe are required to play key role in developing overall conducive eco-system for student fraternity. Some previous studies have analysed student entrepreneurship and impact of entrepreneurship courses. However, role of universities as provider and enabler of entrepreneurial environment and its impact on entrepreneurial intent among students has not been studied in a cross-cultural context. Considering this, present study seeks to examine role played by universities in fostering entrepreneurial intention among post-graduate students.\nDesign/methodology/approach: Researcher has taken the framework suggested by Kraaijenbrink et al. (2010) to understand university environment and theory of planned behaviour (Ajzen, 1985) to measure entrepreneurial intention and its antecedents among final year post-graduate management students of India, Malaysia and Singapore. Total sample size is 1,097. Data are analysed with help of exploratory factor analysis, multivariate analysis of variance (MANOVA) and structural equation modelling (SEM).\nFindings: Two factors emerge out of analysis in relation to university environment and support: first, targeted cognitive and non-cognitive support and second, general educational support. With help of SEM, attempt is made to find relationship between these two factors and entrepreneurial intention. It is found that university environment and support has significantly positive relationship with perceived behavioural control. With help of MANOVA, it is found that there is statistically significant difference between perceived university environment and support factors among students of India, Singapore and Malaysia. With this, for both factors highest mean score is found among students of Malaysia, followed by students of Singapore and India.\nOriginality/value: Study has closely examined role played by university environment and support to foster entrepreneurship among young students. Findings of the study can be used by post-graduate educational institutes to design pedagogy, create enabling entrepreneurship support system and work towards becoming an entrepreneurial university.\n",
"role": "assistant"
}
] |
Accumulated knowledge and innovation as antecedents of reputation in new ventures
|
[
"Knowledge",
"Reputation",
"Innovation",
"Entrepreneur",
"New ventures"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: One of the most important factors that determine the success of a new venture in many industries is the creation and maintenance of a positive reputation in the market (Fischer and Reuber, 2007). Establishing such a reputation depends on several factors, among which knowledge, innovation, quality, and differentiation are especially important in defining a positive image for a new company.\nReputation in new ventures: Reputation is considered an intangible resource that a firm possesses. A positive organizational reputation is critical because of its potential to create value, but also because its intangible character makes replication by competing firms considerably more difficult (Roberts and Dowling, 2002). Therefore, it is a competitive tool and organizations should not downplay its importance. In fact, a strong reputation is a powerful element in the search for competitive advantage and competitiveness (Gjerde and Slotnick, 2004; Bergh et al., 2010; Aula and Mantere, 2013).\nAntecedents of reputation in new ventures from the knowledge-based view: According to Hall (1993), due to its importance reputation should receive constant management attention, and in fact activities related to forming and maintaining a good reputation are key tasks in setting up a new venture (Fischer and Reuber, 2007). The formation of the organizational reputation entails a dynamic process where some elements can determine its main characteristics.\nMethod: Sample and data collection\nResults: Analysis of the factors\nDiscussion and conclusions: This work attempts to shed some light on Abimbola and Vallaster's (2007) call for more research on reputation-based sources that connect the small and medium enterprises with their customers. Although there have been some works which have addressed reputation from an internal perspective in the context of new firms (e.g. Goldberg et al., 2003) or even before the creation for its specific role on capital search (e.g. Shane and Cable, 2002), to our knowledge this work is the first one to empirically test the impact of knowledge-based factors on reputation in new ventures, once the firm is already on the market. Thus, this study contributes to the identification of internal strategies to build an organizational reputation and overcome the traditional liability of newness (Stinchcombe, 1965) linked to these young organizations. As new firms have a limited accumulated production track record to show their reputational traits to the market, the search for strategies and assets become fundamental and the literature has been scarce in providing them (e.g. Goldberg et al., 2003). Moreover, despite the literature on general management and the above hinted studies which show the relevance of the entrepreneur's reputation to set up a new venture, the interest for conducting empirical analysis of the direct effect of reputation in new firms' competitiveness, has not been corresponded to the same extent, especially in non-high-tech industries.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the effect of an entrepreneur's accumulated knowledge and firm's innovation on the development of reputation in the early years of a new venture from a knowledge-based approach.\nDesign/methodology/approach: The study proposes a model that is tested with a sample of 130 firms in non-high-tech industries from the Canary Islands (Spain) using structural equation modelling. Data were collected through a survey.\nFindings: This study provide interesting insights on the effect of reputation on the performance in new ventures, along with antecedents of the new firm's reputation from the knowledge-based view. The findings confirm that innovation and prior knowledge play important roles in the development of reputation in the early years of a new venture and that reputation has a significant effect on the performance of a new firm. The entrepreneur's stock of knowledge does not reveal itself as a significant determinant of innovation and knowledge creation in this context.\nResearch limitations/implications: The entrepreneur's accumulated knowledge should be seen as a valuable existing asset for a new venture, and innovation and knowledge creation can be used to develop core competencies in orientating the strategic direction of a new venture. Both elements become fundamental despite addressing non-high-tech industries.\nPractical implications: Entrepreneurs should be aware of the key role that the creation and the stock of knowledge play in the first years of company life and this research shows how significant this relationship with the initial reputation and performance of new venture in non-high-technology industries is.\nOriginality/value: There is a relative scarcity of studies on reputation-building strategies in new entrepreneurial ventures, and the present study adopts an original knowledge-based perspective to shed new light on the analysis of reputation.\n",
"role": "assistant"
}
] |
New paradigm for patterns of home packaged food intake
|
[
"Food packaging",
"Assessment",
"Portugal",
"Consumers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Packaging plays an essential role in the food supply protecting and containing food from processing and manufacturing, through distribution, handling and storage to the final consumer. Without packaging, food distribution would be a messy, inefficient and much more costly exercise. The packaging sector represents about 2 per cent of the GNP in developed countries and about half of all packaging is dedicated to food (Robertson, 2006). Packaging functions may be described as:* protection and containment;* preservation;* information; and* convenience and service.These basic functions are, directly or indirectly, important for the physical, chemical and microbiological safety of foods. For the majority of food products, the protection afforded by the package is an essential part of the preservation process and the requirements for a packaging system intended for a fresh, frozen, dehydrated, thermal or aseptic processed product, in terms of oxygen, moisture and light barrier, etc. are all different.Despite the important role it plays, packaging is, in the view of many consumers, a waste of resources, which ends up as an environmental burden. Such views arise because the functions which packaging has to perform are either unknown or not fully considered and appreciated by the consumer (Robertson, 2006). When packaging functions properly, it tends to be unnoticed and taken for granted. Packaging is most likely to be noticed when the bag stretches instead of opening, when the easy-open feature fails or when the child-resistant feature is impossible to decipher by anyone other than a child (Downes, 1989).Departing from the basic functions of protection and preservation, packaging systems have evolved to more added value functions related to the function and use of the product. Packaging is also now a major source of information concerning the product, relevant at both the distribution and consumer levels. Traceability, which is key to modern concepts of food control, is unthinkable without the information represented on packaging. When products are being designed or altered, packaging-based details are key in assuring quality and safety, convenience of sizes and formats, service and support of busy modern lifestyles and environmental impact minimization. Of the convenience aspects demanded by the modern consumer, easy opening, reclosing between uses, dosing features, cook-in and oven safe packages are examples of how packaging can respond.Concerning food safety and packaging, the consumer is frequently misled by multiple media, scientifically unsound messages from dubious sources, often appearing as messages such as \"Plastics leach chemical contaminants in the microwave or freezer at levels that are harmful to health\" or \"Plastics contain dioxins\". It is recognised that that human exposure to chemicals from packaging can occur as a result of migration from the packaging materials into foodstuffs. It is right that this should be controlled to an extent which reflects the human-health risk represented, but alarmist reports cause concerns among consumers, who are largely unaware of the stringent assessment and approval processes which packages must pass, by authorities such as EFSA or FDA, before they can be used commercially.Exposure, in a dietary context, is defined as the amount of a certain substance that is consumed (Holmes et al., 2005) and is usually expressed as the amount of substance per mass of consumer body weight per day. In packaging applications, exposure can be expressed in terms of the amount of packaging used, which contains the chemical in question and the migration of the chemical into the food, see equation (1). For assessing the exposure to packaging chemicals, it is necessary to know what type of food (chemical and physical nature) is packaged in what type of material, as this determines the presence and concentration of the chemical and influences the potential for migration into the type of food (Pocas and Hogg, 2007): Equation 1 The Migration term of equation (1) represents the amount of substance migrating from the package into the contained food. Migration data may be obtained from monitoring levels of chemicals in real food systems, or more commonly from migration experiments, carried out under controlled conditions of time and temperature of contact between the materials and a food simulant instead of the food itself. Mathematical models may also be used to simulate migration data (Pocas et al., 2008).The Packaging usage term represents the amount of packaging used to pack the food that the consumer eats. Packaging usage data are, unfortunately, very scarce, with most consumption surveys designed to gather information on the food itself regardless of the packaging system containing it. Adequate information of packaging usage, usable to assess exposure to migrants, is thus not readily available. Data on the most aspects of food packaging in Portugal are very scarce. A report with market data collected through statistics of national and trade associations and through interviews with major food companies and packaging suppliers was published in 1995 (Pocas and Xara, 1995). This publication, apart from being out of date, does not have the detail required for any relevant exposure study. One of the objectives of a nationally (Portuguese) funded research project MIGRAMODEL, was the development of a framework for exposure assessment of packaging migrants. This framework includes packaging usage data collected at consumer level, parts of which are presented in this current paper. The collection of updated and more complete data on packaging usage will enable the refinement and possible correction of current EU assumptions used for exposure assessments of packaging migrants.\n2. Methodology: Packages from commercial food products used at home by a sample of consumers during seven days were collected and analyzed in the laboratory. The sample elements were members of MIGRAMODEL project team, relatives and faculty members. The sample, was composed by 62 households, accounting for a total of 205 persons. Separate collection of those packages used by children under 12 was requested. A total of 2,877 packaging items were analysed in this study.Each household was characterized through a short questionnaire, considering: number of persons, weight, age and gender of each member, and the annual aggregate income.The packaging materials and systems used (consumed) by each household were brought into the laboratory and analyzed for:* Type of package, material and geometry: materials were segmented as: glass, tinplate, aluminium, all-plastic multilayer, multilayer (multi-material), paper, coated paper, carton board and plastics. These latter were further segmented into different major types of plastics: PET, PP, HDPE, LDPE, PS. All materials were in direct contact with the food products, except cardboard, which was in all samples, a non-contact material. However, because migration can also occur from non-contact materials, cardboard was also considered in the consumption factors estimation.* Capacity (amount of product), weight and surface area in contact with food of package.* Type of foods: the foods were classified into the following types - aqueous (A), acidic (B), alcoholic (C), fatty (D), dairy products (E) or dry non-fatty foods (None). These correspond to the simulants foreseen in the European food contact legislation. In the case of food items requiring more than one testing simulant, the fattier simulant one was considered.The data were used to estimate:* the global amount of used packaging material and consumed food;* the usage of package as a function of body weight;* the estimates of consumption factors (CF) that represent the fraction of the daily diet expected to contact specific packaging materials (USFDA, 2007); and* the estimates of food type distribution factors (FDF) that reflect the fraction of all food contacting each material that is aqueous, acidic, alcoholic or fatty (USFDA, 2007).Food taken out was not accounted for in this analysis, which obviously contributes to underestimation of the absolute values. Thus, results must be seen as representing the relative quantities in terms of food, not the absolute quantities.\n3. Findings: Characterization of the consumer sample\n4. Implications, limitations and conclusions: The data collected are the base for a more realistic approach for safety and exposure assessment of food packages, in what concerns migration of substances from the package into the food. The amount of packaging (in terms of surface area) that the consumer uses, can be combined with migration data for a certain and specific substance that is present in the packaging material in question, and used in equation (1) to estimate the exposure. The seven-day period of packaging collection used is the period often used in food consumption surveys. However, it is considered insufficient for packaging surveys purpose. One week is probably not enough to consume a whole bottle of olive oil or a large family size of a margarine tub, for example. Therefore, if these items are not ended during the week the survey occurred, they will not be accounted for in the data. Future studies would need to consider a longer period of packaging data collection. An extended period of one month of collection, for example, may allow for more precise results in packaging surveys. The other major limitations to conclusions may be the size of the sample. Nevertheless, in spite of the present work intended to be a pilot study for setting up procedures and tools, the number of packaging items handled and analysed was considerably high in comparison with other complete studies (FSA, 2006). This benefits the representativeness of the data in terms of variability of packaging types and different parameters analysed.The results achieved contribute to a refinement and possible correction of current EU assumptions for exposure assessments when applied to Portuguese population. The data collected on consumption and food-type distribution factors allow the adoption of an approach more refined than the worst-case estimates that can combine a probabilistic approach to model variables for which the distribution of values are known.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to describe a pilot study characterizing the domestic usage of food packaging (amount and type) at the household level.\nDesign/methodology/approach: - Collection and detailed characterization is carried out, from a sample of Portuguese consumers, of packages used at home during a seven-day period.\nFindings: - The daily intake of domestically consumed, packaged food ranged from 5 to 50 g/kgbw. The average intake value for children below 12 years old was 26 g/kgbw. Considering all packaging materials, package usage ranged from 0.1dm2/day.kgbw to 0.6dm2/day.kgbw. Packaging usage factors and food-type distribution factors for each packaging material were estimated.\nResearch limitations/implications: - The period of packaging collection and the size of the sample limit the degree of generalized conclusions that can be drawn.\nPractical implications: - The data collected are the base for the development of a framework and could make an important contribution to the assessment of consumer exposure to substances migrating from packages into food, in this case of the Portuguese consumer.\nOriginality/value: - The results achieved are analysed in relation to the current assumptions made for the safety assessment of materials in the relevant European legislation. Furthermore, the results achieved can be used in a probabilistic approach to these assessments as the distributions of values, and not only the per capita values, are known.\n",
"role": "assistant"
}
] |
Ethno-nationalism and impediments to cooperation in tourism in a post-settlement Cyprus?
|
[
"Governance",
"Politics",
"Hotels",
"Futures"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Cyprus is a politically divided island that consists of two major political entities, the \"Turkish Republic of Northern Cyprus\" and the Republic of Cyprus. The \"Turkish Republic of Northern Cyprus\" is recognized only by Turkey and is populated by Turkish Cypriots and settlers from mainland Turkey; the Republic of Cyprus is a member country of the European Union and is populated mainly by Greek Cypriots. The two have been separated by UN-administered \"Green Line\" since the Turkish invasion of 1974. About a third of the island - in the north - is occupied by Turkey. Here, we discuss how tourism professionals envision the prospects of current and future cooperation between the two prevailing ethnic groups on the island that has a history of bloody interactions and has largely been separated from each other since 1974.\nLiterature review: There is a vast literature linked with tourism and conflict, and recently two major edited works have been produced (Moufakkir and Kelly, 2010; Wohlmuther and Wintersteiner, 2014). There are many other writings that have dealt with the issue of tourism and conflict. Writings focus on specific cases such as Ireland (Anderson and O'Dowd, 1999; Anson, 1999; Boyd, 1999; Leslie, 1999; McDowell, 2008; Zuelow, 2006), the Basque country (Abadie and Gardeazabal, 2003; Plaza, 2002), the Middle East (Brin, 2006; Cohen-Hattab, 2004; Gelbman, 2008; Gelbman and Timothy, 2011; Mansfeld, 1996), or Korea (Seongseop Kim and Prideaux, 2006; Chung and Yu, 2001). But there are other writings that deal with more general/theoretical issues on the topic (Arana and Leon, 2008; Coshall, 2003; Etxebarria and Gomez Uranga, 2002; Hall and O'Sullivan, 1996; Law, 2006; Neumayer, 2004). Webster and Ivanov (2014) make a contribution in which they look into how liberal concepts of economic and social interactions should result in encouragement of economic growth; a virtuous cycle of economic benefit and social encouragement, with economic growth, is created that seems to work in conditions in which the liberal preconditions are met.\nMethodology: Asking questions about the future in tourism is now an established strand in the tourism literature. The leader in the field has been Ian Yeoman (Yeoman and Lederer, 2005; Yeoman, 2008, 2012) but many others have been contributing (Dwyer et al., 2009; Leigh et al., 2013) and this has culminated - in an academic sense - in a new journal, Journal of Tourism Futures. The approach taken in this paper is more constrained than that of most futurologists of tourism, who tend to focus on megatrends to identify the broad sweep of change. Our approach is to focus more narrowly on the likelihood of future cooperation. Given the long history of conflict, the relative dearth of cooperation between tourism professionals North and South has not been surprising. That there has been any cooperation at all (see Webster et al., 2009 and Table I) is remarkable.\nResults and analysis: Survey of tourism professionals - quantitative\nConclusion: The recent revival of negotiations between the elected leaders of Cyprus, North and South, augurs well for the tourism industry. The survey data, and interviews with tourism professionals and representatives, all show overall that there is an expectation on both sides of a win-win for tourism in the event of a settlement acceptable to both sides. This expectation is likely to be realised in the event of a settlement. As discussed above, there are differences between the tourist sectors North and South. These differences may in part explain the different attitudes of tourism professionals North and South. However, the differences are also likely to enhance the attractiveness of Cyprus as a tourist destination in the event of a settlement. Yorucu and Mehmet (2011) go further, focussing on specific countries likely to remain among the important origins of tourists to Cyprus, namely, Turkey, Greece and the UK. They conclude that Turkish and Greek tourists have income elastic demand for Cyprus as a tourist destination. Recent economic problems in Greece will therefore reduce Greek tourist response to a settlement in Cyprus but Turkey will become a more important origin for tourism North and South. UK tourists, being more price sensitive, will become less important for tourism in the North, Yorucu and Mehmet (2011) suggest. As a policy recommendation they suggest for the North, but this could be echoed for some tourism providers in the South as well, a focus on more high-income, special interest tourism products, reducing the dependence on low-price, mass market tourism products.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to clarify the position of Greek Cypriot and Turkish Cypriot professionals in the hotel and tourism industry on the island of Cyprus with regards to their expectations regarding the benefit of a political solution to the Cyprus problem on the island.\nDesign/methodology/approach: The paper uses data from two surveys in both political entities of the island. One survey is a survey of hotel owners, managers of hotels and travel agencies in both political entities on the island. The other is semi-structured interview with leading professionals in the hospitality and tourism industry in both political entities.\nFindings: The surveys indicate that there is an expectation from professionals in both entities that tourism will benefit all following a solution, with large increases in incoming tourism to Cyprus.\nPractical implications: The findings indicate that there are substantial expectations that there will be benefits for all following a solution to the Cyprus problem.\nOriginality/value: This is a first future-oriented paper regarding the expectations of major players in the hotel and tourism industry in both political entities on the island.\n",
"role": "assistant"
}
] |
Teamwork makes the team work: An interview with Dr Meredith Belbin
|
[
"Human resources",
"Human resource management",
"Management teams",
"Team dynamics",
"Team roles",
"Organizational structure",
"Testing",
"Team working"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: In order to get the most from a management team an understanding of the people involved, and how they interact, is essential. If the mix is not right team effectiveness may be less than the sum of its individual parts.\nWith the technological advancements of recent years, team members need no longer be based in the same office, or indeed country. How does geographical dispersion affect team dynamics?: Geographical dispersion does not help teamwork. Emails do not ideally serve interpersonal communication. Gesture, facial expression, inflexion of voice and the chance of getting quick answers to questions; these are all missing and so cause misunderstanding. Such shortcomings may account for the rise in video conferences - a trend that should be welcomed.\nIn your opinion, what are the biggest issues facing human resource managers today and how can they be overcome?: HR is often regarded by line management as a resource (akin to other resources like finance and materials) and by accountants as a cost. Personnel Management, its forerunner, maintained a focus on people issues. Yet I am often impressed with the depth of understanding I find among some of the HR trainees that pass through our office. Such assets are wasted when decisions depend solely on line management with a superficial knowledge of the underlying people issues. Our response to this problem will be to run joint workshops for line managers and HR personnel, as planned for 2013.\nManagement thinker Gary Hamel said in his recent CiPD keynote that management, in its current form, is a \"busted flush\" - that the practices developed 50-60 years ago (which are still prevalent today) are no longer fit for purpose. How do you see management, and the teaching of management, evolving?: Gary Hamel is right in principle. The key to this \"busted flush\", in my view, is the focus on leadership. Whatever it should mean, it conveys most the notion of a top leader who singularly shows the way to others. I consider this concept lies at the root of the banking crisis and the highly irrational decisions taken by ambitious individualists.\nYou have been involved in the study of organizations and the people within them for many years. What has been the most significant development over that period of time? And what do you see being most significant in the coming years?: The biggest change over the years has been the rise in the entry of those with higher education, including women, into the senior echelons of management. Yet the age-old concepts of management have barely shifted. It is this paradox that will underlie the upheavals likely to happen in the next few years.\nWhat are you currently working on?: The new focus of our office in 2013 will be Career Development within the Organization, a matter on which I take a keen interest. As far as my personal plans are concerned, however, I intend to reinvent myself as a futureologist. I laid the foundations for this in my most recent book Managing Genetic Diversity. The goal now is to move away from the intellectual roots of this book and to explain its relevance to the everyday problems of the modern world.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to provide an interview with Dr Meredith Belbin.\nDesign/methodology/approach: - The paper provides an interview with Dr Meredith Belbin.\nFindings: - The paper explores Belbin's thoughts on current issues including HR management, organizational structure and geographical dispersion.\nOriginality/value: - The interview offers insight for HR professionals from the founder of Team Roles theory.\n",
"role": "assistant"
}
] |
Cluster analysis in empirical OM research: survey and recommendations
|
[
"Empirical",
"Literature review",
"Methodology for operations management"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Clustering methods form groups or clusters of objects based on some measure of similarity of attributes between those objects. In a typical operations management (OM) application, these objects can be manufacturing facilities, business units, firms, or managers; while the attributes are variables such as firm design flexibility, business size, or tenure of managerial experience. Similarity between these objects is commonly expressed in a measure of proximity represented quantitatively by the squared Euclidean distances between pairs of objects based on the set of variables obtained from single- or multi-item scales. There are a number of emergent clustering techniques that might be viable for some applications (see Brusco et al., 2012), which include: mixture-model clustering (McLachlan and Peel, 2000), p-median clustering (Kohn et al., 2010), and neural clustering (Du, 2010). Nevertheless, the most popular clustering methods within OM research have been Ward's (1963) hierarchical method and the K-means, nonhierarchical (Steinhaus, 1956) method, both of which seek to minimize the sum of the squared Euclidean distances between objects and their cluster centroids. Hereafter, we refer to this sum-of-squares criterion as within-cluster sum-of-squared error (SSE).\nRecommended approaches for validation: Table I provides a summary of our validation recommendations as a sequence of steps in the clustering process: the use of variable selection methods to minimize or remove the influence of masking variables, properly standardizing clustering variables when necessary, use of multiple restarts of the K-means algorithm to mitigate possible effects from alternative local optima, selecting the number of clusters using salient objective criteria, using internal consistency measures to assess the robustness of the clustering solution among subsamples, and examining cluster differences on variables exogenous to the clustering process.\nCoding study: Journals\nDiscussion and conclusions: Recommendations for the practice of cluster analysis in OM\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is twofold. First, the authors provide a survey of operations management (OM) research applications of traditional hierarchical and nonhierarchical clustering methods with respect to key decisions that are central to a valid analysis. Second, the authors offer recommendations for practice with respect to these decisions.\nDesign/methodology/approach: A coding study was conducted for 97 cluster analyses reported in six OM journals during the period spanning 1994-2015. Data were collected with respect to: variable selection, variable standardization, method, selection of the number of clusters, consistency/stability of the clustering solution, and profiling of the clusters based on exogenous variables. Recommended practices for validation of clustering solutions are provided within the context of this framework.\nFindings: There is considerable variability across clustering applications with respect to the components of validation, as well as a mix of productive and undesirable practices. This justifies the importance of the authors' provision of a schema for conducting a cluster analysis.\nResearch limitations/implications: Certain aspects of the coding study required some degree of subjectivity with respect to interpretation or classification. However, in light of the sheer magnitude of the coding study (97 articles), the authors are confident that an accurate picture of empirical OM clustering applications has been presented.\nPractical implications: The paper provides a critique and synthesis of the practice of cluster analysis in OM research. The coding study provides a thorough foundation for how the key decisions of a cluster analysis have been previously handled in the literature. Both researchers and practitioners are provided with guidelines for performing a valid cluster analysis.\nOriginality/value: To the best of the authors' knowledge, no study of this type has been reported in the OM literature. The authors' recommendations for cluster validation draw from recent studies in other disciplines that are apt to be unfamiliar to many OM researchers.\n",
"role": "assistant"
}
] |
Work design for different generational cohorts : Determining common and idiosyncratic job characteristics
|
[
"Croatia",
"Work design",
"Job characteristics",
"Generational differences",
"Workforce generations"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Workforce differentiation is gaining momentum. Organizations are facing a more diverse set of employees (distinguished by age, gender, race, educational level, occupation, etc.), and therefore they strive to find adequate human resource management (HRM) policies. Workforce demographics, especially generational differences, have become a compelling research topic (e.g. Zemke et al., 2000; Smola and Sutton, 2002; Glass, 2007; Twenge et al., 2010; Joshi et al., 2011; Truxillo and Fraccaroli, 2013).\nWork design and generational cohorts: Work design theory and research have largely overlooked workforce diversity issues. However, increasingly diversified workforce in general and generational differences specifically should encourage scholars and HR professionals to question traditional notions of what makes workers thrive in work contexts. Demographic trends, accompanied with economic, technological and cultural changes, require commensurate shifts in how work is structured and organized (e.g. Egri and Ralston, 2004; Cartwright and Holmes, 2006; Hewlett et al., 2009; Cordery and Parker, 2012). Consequently, work design should meet personal values and work preferences of different generations of employees.\nResearch methodology: Participants and procedure\nResults: Common and idiosyncratic job characteristics among multigenerational workforce\nDiscussion and conclusion: Research implications\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to uncover the nature of job characteristics related to different generational cohorts (Baby-boomers, Generation X and Generation Y). Significant differences between four task and four social job characteristics across generational cohorts have been revealed.\nDesign/methodology/approach: The empirical research was conducted through a field study of employees from large-sized Croatian organizations. A cross-sectional and cross-occupational research design was applied. A total of 512 knowledge workers (139 managers and 373 professionals) participated in the research. Descriptive and inferential statistical methods were used to determine and compare work design across generations.\nFindings: The results indicate that job characteristics are not equally represented within different generational cohorts. While the nature of task job characteristics is mostly irrespective of generations, social job characteristics to some extent differ among generational cohorts. High task variety, reasonably high task identity, and a moderate level of both received interdependence and task significance are recognized as common job characteristics of knowledge workers across generations. However, jobs of Baby-boomers, Xers, and Yers are idiosyncratic for work autonomy, interaction with others, initiated interdependence, and teamwork. Additionally, the inclusion of the work type as a control variable revealed that interaction with others does differ but only among generations of professionals.\nOriginality/value: The present study is the first research in which generational similarities and differences have been empirically examined through job characteristics. The authors focused on knowledge workers within an under-researched context (studies about knowledge workers, work design and generational differences are rare or non-existent in south-eastern European countries), making this systematic investigation unique and practically significant.\n",
"role": "assistant"
}
] |
Futures studies and public decision making in Sweden
|
[
"Sweden",
"Futures markets",
"Decision making",
"Strategic planning"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Sweden is a constitutional monarchy with a parliamentary system of government (www.sweden.se). The government works through centralised government authorities, its reach being extended through local offices (21 counties and 290 municipalities) (www.regeringen.se).\n1. The players involved in futures studies: Futures-oriented works are mainly carried out at central level but also at regional and municipal levels. Projects are frequently characterised by a public/ private partnership and big companies often have an active futures studies department.There are many, varied players involved in futures studies and they work interactively. Some of the most important ones, at national level, are presented below, starting with the institutions closest to decision-making powers.Over and above the planning and prospective units in certain ministries, the government agencies which are responsible for advisory work in a given field frequently include a futures-oriented dimension in their work. This is the case, for example, of the National Institute for Working Life, which reports to the Ministry of Industry, Employment and Communications or the National Board of Housing, Building and Planning which reports to the Ministry of the Environment.It is common practice for the government to appoint a commission of inquiry tasked with conducting a detailed study of a matter in order to then propose changes in the law or reforms (see Box 1 on the Lindbeck Commission). This examination can be entrusted to one person, but generally speaking, in addition to its chairperson, a commission of inquiry includes members of the political parties represented in Parliament as well as representatives from various organisations and experts from the research world. Discussing the commissions' reports with private agencies and organisations is an essential part of Swedish administrative practice.The Institute for Futures Studies (Framtidsstudier) (see Box 2) was set up in 1987 to replace the Secretariat for Futures Studies set up in 1973, which initially functioned as a complement to the Prime Minister's Office and then as a unit of the Swedish Council for Planning and Co-ordination of research. Today this is an independent institute, attached to the Ministry of Education and Research, which deals with developments in society (www.framtidsstudier.se).Trade unions, political parties, professional organisations, NGOs and other associations are very much involved in political discussion and public debate in Sweden. They take part in prospective thinking through their involvement in the commissions of inquiry or, autonomously, by putting forward their own analyses.Researchers participate in the public decision process through their expertise called upon for diagnosis, evaluation and futures studies. They take part in the futures-oriented work done by commissions of inquiry or government agencies. Furthermore, they are greatly involved in scientific and technology foresights which are initiated by academies of science and research councils (see Box 3) (www.tekniskframsyn.nu).\n2. Relations with an independent stamp: The relationship between futures studies and public decision in Sweden is of many forms. It can be direct in the case of commissions of inquiry. These are officially set up by the government and their results are directly put into use when working out public policies.Apart from the commissions, futures-oriented studies are not generally commissioned by the government, even though the latter may provide support for certain initiatives (funding, legitimacy). For instance, the idea for the first technology foresight came from the Royal Swedish Academy of Engineering Sciences. However, the project designers would not have launched it without the government's backing, which guaranteed that the results of the exercise would be put into use by the authorities.Relations with government or with Parliament increase during the phase of communicating the results of the work and the government has no influence on the way projects are run and the chosen themes for debate. Likewise, there is a pretty clear division of labour between the prospective debate on the one hand, which consists in identifying factors of change and future challenges (and for which experts are greatly called upon), and the task of drafting public policies on the other hand, conducted by the government and ministries.So, the relationship between futures studies and public decision is stamped by the principle of independence. This principle is behind the change of status, during the 1980s, of the Secretariat for long-term studies, directly subordinate to the Prime Minister, which became the Institute for Futures Studies (an independent research foundation). According to Joakim Palme, director of the Institute, this change was brought about by the principle of political independence and by the need to thwart the overly technocratic direction that futures-oriented analyses had taken in the 1970s. The Lindbeck Commission is a further illustration of the strength of the principle of independence in expertise in Sweden. Although the Commission was set up by the government, Assar Lindbeck, its chairman, was in a position to demand and get total freedom in choosing experts, the work method, the themes dealt with and the conclusions. Moreover, the government only read the conclusions of the report on the day of the press conference.Generally speaking, it seems that the relationship of independence between futures studies (and to a greater extent, advisory work) and public decision is pretty exemplary in that it does not hamper interplay. Researchers, although attached to their independence, want to be heard by the authorities (advisory work is clearly a part of their missions) and, with this aim in view, have to take up issues on the political agenda. The authorities, for their part, need the researchers' expertise to devise public policies. The overall coherence is, moreover, strengthened by the fact that researchers and decision-makers are educated in the same universities.\n3. Futures studies: a tool for designing public policies: The independence between futures study and public decision results in relations that are more often than not indirect and informal. These relations never-the-less have a real impact, as futures study is a genuine public policy instrument in Sweden. In this respect, it essentially fulfils three aims:1. N: Futures-oriented analyses contribute to public debate and to making Swedes aware of future challenges.2. Futures study is a frequently used method for consultation which takes place prior to major social reforms.3. Futures study, in particular through technology and scientific foresight projects, is a tool to define strategic priorities.Raising awareness and taking part in public debate\n4. Conclusion: look ahead more in order to avoid crises: Although stamped by independence with regard to political powers, futures study in Sweden is a decisive part of the public decision process. However, the players involved in futures studies as in public decision are faced with a dilemma it is hard to get out of. The success of futures-oriented studies, among the general public or in terms of impact on the decision, is even greater because they are carried out in a context of crisis or indeed because they focus on themes that are already very much present in the public debate. Consequently, Assar Lindbeck considers that the success of the Commission on Sweden's future is mainly due to the context of crisis which shook the Swedes on account of its brutality and depth. Likewise, the effect of the first technology foresight lies very much in the culture of innovation that has grown in Sweden these past few years. And yet, futures studies should provide the means to forecast potential crises in order to avoid them or limit their effects. For instance, Elisabeth Backteman and Bjorn Andersson consider that matters relating to the social, cultural and political integration of immigrants are today the main theme missing in Swedish futures studies. The effects of population ageing and EU enlargement will necessarily be an influx of immigrants. It therefore seems urgent to think ahead about the main challenges of a multicultural society in order to avoid a potential increase in xenophobic ideas.\n",
"role": "user"
},
{
"content": "Purpose: - This paper seeks to explore the interrelationships between futures studies and public decision making in Sweden.\nDesign/methodology/approach: - Based on a series of interviews conducted in Stockholm in December 2003, this paper presents the main players involved in futures studies in Sweden. It focuses especially on the activities of the Institute for Futures Studies, and that of the Lindbeck Commission on the Future of Sweden.\nFindings: - Futures studies are a decisive part of the public decision process in Sweden, where they appeared after the Second World War and gained momentum from the 1970s onwards. The paper gives a detailed account of the interplay between the various actors. There is a strong independence between the public decision-making system and the producers of futures studies. Nevertheless, futures studies prove an important instrument for public policy in Sweden, as they raise awareness in the public debate, serve as a tool for consultation, and help define scientific and technological priorities.\nOriginality/value: - Provides an insight into the state of futures research in Sweden.\n",
"role": "assistant"
}
] |
Drivers of franchisor growth: a meta-analysis
|
[
"Franchising",
"Meta-analysis",
"Franchisor growth",
"Geographic dispersion",
"Outlet growth"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Franchising is a very important marketing channel, and in the USA alone, it generates $892bn sales in 2015 (IHS Economics, 2016) and contributes about 30 per cent of US retail sales (Zeidman, 2010). There are over 78,000 franchise units in Canada and 44,200 franchise units in the UK by 2015 (British Franchise Association, 2016; Canadian Franchise Association, 2016). In emerging economies such as China and Brazil, the growth of franchises has been substantial (International Trade Administration, 2016). As one of the fastest growing marketing channels, franchising has attracted research attention from diverse disciplines such as law (Agrawal and Lal, 1995), economics (Lafontaine and Slade, 1997), management (Combs and Ketchen, 2003; Combs et al., 2011) and marketing (Kaufmann and Dant, 1998; Perry et al., 2002; Harmon and Griffiths, 2008; Bordonaba-Juste et al., 2011).\n2. Hypothesis development: Franchisor growth has been measured in the franchise literature in a number of ways:\n3. Methodology: 3.1 Collection and coding of franchising studies\n4. Results: 4.1 Bivariate correlation analysis\n5. Discussion: This meta-analysis examines the relationship between franchisor growth and its key drivers and several noteworthy observations emerge from the study. Our study confirms the diverse nature of the relationship between age and the various measures of growth, as we find that age relates positively to geographic dispersion but negatively to the number of new outlets and outlet growth rate. The negative relationships that we find between age and outlet growth rate and also number of new units provides support to the resource scarcity theorists who suggest that younger franchisors grow aggressively during their early stages to acquire the financial and human resources that they need to compete with larger competitors. Although critics of resource scarcity theory (Rubin, 1978) suggest that the risks, costs and inefficiencies associated with franchising are so high that young franchisees would be reluctant to use franchising as an early tool for growth, our results instead confirm that firms are more likely to franchise earlier in their lifecycle than later (Castrogiovanni et al., 2006; Combs and Ketchen, 1999). Size also has varying relationships with the different growth measures: It relates positively to geographic dispersion and number of new outlets but negatively to outlet growth rate. The positive relationship between size and geographic dispersion and also number of new outlets, reflects the financial and managerial resource capabilities of larger franchises to operate successfully in multiple geographic locations and add more outlets. However, size is negatively related to outlet growth rate, perhaps because of mathematical properties, in that larger franchises need to add many more units to reach the same growth rate as smaller franchise systems.\n6. Limitations: As does any methodological approach, a meta-analysis suffers inherent limitations. The constructs we use are limited to those with evidence of four or more effects in prior research. Thus, our study offers an examination of the most commonly studied drivers and not a complete review of all potentially important variables (Riggle et al., 2009). Many of the studies we identified did not report correlation coefficients (r), and our efforts to retrieve these correlations were not always successful. Therefore, we used formulas provided by Rosenthal (1994) and Peterson and Brown (2005) to derive Pearson's product moment correlations. The relationships between correlations calculated using these formulae and the original correlations are high (r = 0.75) (Peterson and Brown, 2005), but it would have been preferable if we could have obtained and used the original correlations.\n",
"role": "user"
},
{
"content": "Purpose: This study aims to help resolve some of the inconsistencies of the relationships between franchisor growth and its drivers in prior literature.\nDesign/methodology/approach: First, this study provides a meta-analysis with bivariate correlation analysis and moderation analysis. It then offers an additional analysis of secondary data to shed further light on the relationship between franchisor growth and its drivers.\nFindings: This study confirms the diverse nature of the relationship between the various measures of growth and drivers. It finds that proportion of outlets franchised and brand reputation have the strongest relationships with geographic dispersion; age and proportion of outlets franchised have the strongest relationships with outlet growth rate; and size has the strongest relationship with the number of new outlets. In addition, these multiple relationships are moderated by all three research characteristics that this study investigates, including data source, time frame and industry context.\nResearch limitations/implications: This meta-analysis merely offers an examination of the most commonly studied drivers and not a complete review of all potentially important variables. It calls for further research that examines the factors that lead to franchisor growth and performance in general.\nPractical implications: Managers of young franchisors do not need to rush to expand their business across a wide range of geographic regions. Young franchisors instead should focus initially on gaining maturity, developing their business concept, building an attractive track record and improving their brand reputation. Beyond a strong brand and well-developed business concept, franchisors can attract potential franchisees by reassuring them and making them feel secure about their investment.\nOriginality/value: This study includes a bivariate analysis that was used to conduct a meta-analysis and also an empirical analysis of secondary data. By conducting the secondary data analysis, we were able to examine the extent to which the meta-analysis results of this study could be extended beyond the time period for papers included in the meta-analysis.\n",
"role": "assistant"
}
] |
Interactive foresight on the Hungarian SMEs
|
[
"Foresight",
"Forecasting",
"Research methods",
"Scenario planning",
"Knowledge creation",
"Regional development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Foresight processes are applied very often in the futures fields to shape futures ideas concerning certain fields of reality with the participation of futurists, experts and/or stakeholders and lay people. Different processes are used depending on the aims, the subjects and the tasks of foresight as well as the needs of customers. We agree with the statement that basically there are two different kinds of foresight meanings at present (Hideg, 2007). One of them is the foresight in futures studies which emphasizes the development of future intelligence or future literacy of participants (Slaughter, 1995, 2004, 2008; Hideg and Novaky, 1998; Hideg et al., 2013; Inayatullah, 2006; Stevenson, 2006; Miller, 2007). The other one is the praxis foresight which is developed to support decision making focusing mainly on providing future-oriented recommendations to policy makers and business men (Miles, 2008; Havas et al., 2010; European Foresight Monitoring Network, 2011).\n2. Theoretical and methodological considerations: Interactive foresight is a kind of scientifically based practical foresight activity that updates the practice of participatory futures based on the theoretical considerations of co-evolutionary paradigm of integral futures (Hideg, 2013). The idea of integral futures was first shaped by R. Slaughter. His concept of integral futures means the further development of critical futures in which different kinds of knowledge integration could be achieved by development of personal intelligence of participants (Slaughter, 2008). We agree with Slaughter from that point of view that the development of personality is also carried out in an integral futures activity, but shared and socially accepted futures ideas which are underpinned by scientific and experimental approaches and knowledge and emerged by interactions among different creative participants also have to be shaped in parallel with it. Without the latter, the integral futures cannot be interpreted. Therefore we differentiate between the integral futures as a scientific field and the complex future ideas about certain fields of reality that are the results of participatory foresight activities (Hideg, 2013). According to our concept, the theoretical integral futures could comply with the co-evolutionary paradigm. It means that the theoretical integral futures should be focused on the subject domain of futures and give interpretation and methodological consideration of future as multiple development paths of open and co-evolutionarily changing complex systems. The co-evolutionary interpretation of system changing supposes that complex system is working in mutual interaction with its environment that consists of other complex and evolving systems. Complex system that can be subject domain of futures includes activity and future ideas of human agency as a component system that also interacts within the complex system that is under discussion. This interpretation or comprehension of future makes possible to compose a research perspective in which integration of knowledge acquisition or different kinds of knowledge about society can take place on the basis of complex system dynamics.\n3. Emerging visions of SMEs in the CHR: 3.1 About CHR\n4. Conclusion and discussion: This foresight exercise has proven that multiple interactivity can be built up into the integral foresight activity and sustained during its whole process. The online presence can automatically ensure to continue or restart the foresight activity at any time. The suitably constructed jigsaw puzzle can make quantitatively measurable the participants' learning process. Unfortunately, amid circumstances, the majority of potential stakeholders could not overcome its fears of the present and the future and remained passive and did not take part in the foresight exercise. At the average level of future orientation, the active participants could shape alternative futures of CHR and dialogue and evaluate them. The futurists could build participants' alternative futures for CHR into the \"Hungary - 2025\" social scenarios. It presents that the regional/meso and the national/macro levels of possible futures can be interconnected in an integral foresight exercise.\n",
"role": "user"
},
{
"content": "Purpose: The aim of this study is to present a concept of interactive foresight process, its theoretical and methodological considerations and a foresight exercise concerning the development of knowledge economy in the Central Hungarian Region.\nDesign/methodology/approach: A methodology of interactive foresight process for creating regional future concepts is developed, which is based on a specific meaning of integral futures and uses online solutions, too.\nFindings: Personal meetings with small and medium enterprise (SME) stakeholders and the works of interactive communications with feedbacks within and among stakeholder groups was organized around the research homepage. The networking created the interconnection and the feedbacks between the stakeholders and the futurist group in the process of shaping regional future ideas. The online networking is running.\nResearch limitations/implications: The low number of stakeholders can limit the validity and acceptance of futures ideas created by this process.\nPractical implications: The developed interactive foresight process can also be applicable at different organizational levels and in different fields for shaping shared future ideas.\nSocial implications: Application of interactive foresight process can contribute to the development of anticipatory democracy.\nOriginality/value: A theoretically based interactive foresight process has been developed in which stakeholders can participate not only interactively in the foresight process but they can implement the achievements in their enterprising activity as well. The participants were interested in foresight and cooperative during the whole process because they learned the use of foresight tools through collective solution of practical tasks which were important for them.\n",
"role": "assistant"
}
] |
Continuing professional development needs of teachers in schools for competence-based vocational education: A case study from The Netherlands
|
[
"Continuing professional development",
"Teachers",
"Vocational training",
"The Netherlands",
"Competence based training"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Over the last 15 years, secondary schools in The Netherlands have been granted increasing autonomy by the Dutch central government to reform their curriculum and teaching methods. Many reforms have since been undertaken to make educational programs more competence based, both to decrease dropout rates and to align the Dutch system with the qualification demands set by European Union benchmarks (Ministerie van OC&W, 2005, p. 33). At the same time, schools have been given more autonomy to determine their own human resources (HR) policies, which used to be centrally regulated at the national government level before.\nResearch context: changing education programs in The Netherlands: Before presenting the theoretical background to the case study, this section will provide information about the particular context in which the aforementioned developments are taking place. The Dutch educational system is based on eight years of primary education (ages four-12). Pupils can continue their educational careers taking one of three pathways:1. a four-year prevocational secondary-education program (PVSE, in Dutch: VMBO);2. a five-year senior general secondary-education program (in Dutch: HAVO); and3. a six-year pre-university education program (in Dutch: VWO).Approximately 60 per cent of the children between 12 and 16 years old are enrolled in PVSE, which is the pathway at the focus of the current study. PVSE prepares pupils for an entry into senior secondary vocational education (SSVE), and, eventually, for careers in various vocational contexts (educating them for occupations such as: hair dressing, carpentry, construction, and so forth).The introduction of a competence-based qualification system within SSVE has required PVSE to restructure their programs towards competence-based education. Both PVSE and SSVE have been experimenting with new competence-based educational programs for a number of years now (e.g. De Bruijn, 2004; Van der Sanden et al., 2003). These changes at the program level have had serious implications for the tasks and roles of PVSE teachers. Rather than acting as subject matter experts, teachers nowadays are expected to support their pupils by helping them integrate knowledge, skills, and attitudes into personal competences (Onstenk, 1997, 2001). Teachers currently employed within PVSE have not been educated to fulfill these new roles associated with competence-based vocational education; nor did they experience it themselves. Therefore, they need to engage in CPD activities so as to learn how to better provide competence-based vocational education to pupils. Simultaneously, PVSE schools need to support teachers in making these changes, for example, by creating opportunities for them to experiment with new competence-based education programs. There has been little research that looks at the CPD needs of teachers in schools moving towards competence-based education, which seems to be one of the crucial prerequisites for PVSE schools to come up with an integrated set of HR policies.\nTheoretical background of the study: Competence-based vocational education\nMethods: Context and sample of the study\nResults: Elements of the cognitive apprenticeship model in teachers' action theories\nConclusions and discussion: The aim of the case study presented in this article was to evaluate the extent to which the individual action theories of teachers show elements of two normative frameworks associated with competence-based education: the cognitive apprenticeship model (Collins et al., 1989) and the acquisition vs participation metaphors (Sfard, 1998). For this case study, 12 teachers engaged in an innovation project around the development of a new competence-based education program participated in concept mapping, cued interviews (based on video recordings), semi-structured interviews, and a joint feedback session. The findings of the case study were intended to shed some light on the CPD needs of PVSE teachers in moving towards new roles and tasks associated with competence-based education. Conclusions could help their schools develop better-attuned HR policies to meet teachers' CPD needs.Regarding the cognitive apprenticeship model, teachers seemed to place more emphasis on teaching methods and the sociological context of learning than on learning content and sequencing. These findings might be explained by the focus of the innovation project on the development and implementation of competence-based PVSE. Teachers were probably more focused on teaching methods and sociology because they felt these elements were missing in their daily practice. It is also conceivable that learning content and sequencing elements are simply not as prominent in the individual action theories of current teachers, who were themselves educated by following the sequence and content of textbooks related to their specific subjects.Regarding the acquisition and participation metaphors, a rather mixed picture emerged from the data. On the one hand, the PVSE teachers valued knowledge acquisition by pupils and they believed a structured approach and tight guidance are crucial to ensure the necessary depth and insight. On the other hand, teachers devoted much attention to developing pupils' skills and competencies by having them participate in authentic settings. Those who focused on acquisition experienced the tension of increasingly having to develop their pupils' competencies as well. Teachers who emphasized participation experienced the tension of providing their pupils with enough basic knowledge to be able to function effectively within the community. Those who juxtaposed acquisition and participation saw the potential benefits of both metaphors but experienced tension in having to combine them successfully. And, finally, the few teachers who integrated both metaphors experienced the constant tension of having to negotiate an educational system not yet fully geared to such innovations.The main conclusion from the case study is that no distinctive, crystallized action theories about competence-based PVSE emerged at the individual teacher level; let alone individual teachers' action theories well in line with the normative theory on competence-based education. What does come out is the struggle that individual teachers are going through to get their every-day teaching repertoire more in line with new ideas on competence-based education. The teachers in our sample were all motivated to participate in an innovation project to develop a new competence-based PVSE program; yet, they had serious trouble combining acquisition and participation approaches to pupils' learning as well as balancing the four dimensions of the cognitive apprenticeship model deemed to support competence-based education. There seems to be much scope for schools to develop HR policies that offer teachers CPD activities in this crucial area; especially taking into account that the PVSE school in our sample was relatively far ahead in embracing competence-based education, compared to other schools in The Netherlands.A number of limitations characterize the current study. Data were gathered in one school, representing one school type (PVSE), in one country (The Netherlands); therefore the results cannot be generalized to other contexts without further ado. It seems, however, that if teachers in this progressive school report trouble shifting to their new roles, chances are that many of the more traditional schools will have an even longer change process still ahead (and all CPD needs associated with them). A second limitation is that only qualitative data were gathered in the case study, which again limits the statistical generalizability of the results. On the other hand, four different data sources were used and the low number of respondents involved in this particular innovation project would not have made quantitative data collection useful anyway. Future research should include more cases across several schools and school types, adding larger-scale data collection methods to the ones already used here.The PVSE teachers in our sample did not espouse any distinctive, crystallized action theories about competence-based education, despite the range of different data sources used for the analysis, which gave them many opportunities to put forward any relevant aspects. One explanation could be that not all respondents were able to participate in all four data sources, which is another limitation of our data set.Second, however, one could raise the question to what extent teachers' individual action theories actually predict their performance on the job. Many researchers (e.g. Barnett and Hodson, 2001; Beijaard and Verloop, 1996; Bryan, 2003; Connelly et al., 1997) assume that teachers hold such relatively stable beliefs; however, to what extent their personal theories influence teachers' behavior is open for question. We would argue (with Van der Krogt, 2007) that teachers' behaviors are also influenced by workplace factors (e.g. the classroom setting, their supervisor and colleagues). These contextual cues are possibly more powerful factors influencing teachers' behaviors than their individual action theories are. After class, teachers may take some time to think about the content, teaching methods, sequence, and sociology aspects of teaching in a more reflective way. Further research should focus on the extent to which teachers' behaviors are influenced by their individual action theories relative to other personal and contextual factors.Overall, it would be fair to say that although the individual teachers in our case study referred to several elements of the normative theory on competence-based education (as per the two frameworks), their action theories did not correspond to the normative ideal associated with competence-based education. The two frameworks applied in the present study enable PVSE schools to see in which areas their teachers require additional CPD activities, which could inform much-needed HR policies that many schools are struggling to develop.Teachers are a critical link in implementing an educational reform such as competence-based PVSE (Knapp, 1997). In addition, the quality of education depends to a large extent on the quality of teaching staff, which in turn depends on their participation in CPD activities, at least partly. Managing and maintaining teacher quality, however, is a shared responsibility of the individual and the organization. Integrating continuing teacher development with HR policies in schools therefore represents a promising area for further research. The increased autonomy of schools granted by the Dutch government reinforces the increased interest in HR management and development within school settings. This study has drawn on various disciplines (e.g. educational sciences, HRM, HRD, educational psychology, adult learning), which traditionally have remained mostly separate. It may well be useful for further research to take a more integrative approach of these different literatures.The two models used for our case study underscore this point. Although hailing from the field of educational psychology originally, both of them could help (researchers of) organizations understand what (their) HRD policies are gravitating towards. The participation metaphor described by Sfard (1998) and the sociology dimension of the cognitive apprenticeship model (Collins et al., 1989) are echoed by authors studying the area of learning in organizational contexts (e.g. Billett, 2004; Wenger and Snyder, 2000), while Sfard's acquisition metaphor and the content dimension remain influential in the training and development as well as competence management fields (cf. Sadler-Smith, 2006). It is very likely that HRD/T&D professionals struggling to combine their \"traditional\" knowledge-transfer practices with employees' informal learning (cf. Koornneef et al., 2005) are basically facing the same challenge as the teachers in our case study. Therefore, HRD/T&D professionals could use these models to get a better understanding of effective ways to assess and fulfill employees' learning needs.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this article is to help schools for vocational education determine teachers' continuing professional development needs associated with implementing competence-based education programs, so that these schools can develop better attuned HR policies. It investigates which elements from the cognitive apprenticeship model and from the acquisition and participation metaphors can be identified in the individual action theories of their teachers.\nDesign/methodology/approach: - A case study was conducted in one school for vocational education, where 12 teachers engaged in an innovation project around the development of a new competence-based education program for pupils. They participated in concept mapping, cued interviews (based on video recordings), semi-structured interviews, and a joint feedback session. These four sources were used for an in-depth content analysis of teachers' action theories.\nFindings: - No distinctive, crystallized action theories about competence-based vocational education emerged at the individual teacher level; let alone individual teachers' action theories well in line with the two normative frameworks on competence-based education. The case study shows the struggle that individual teachers are going through to get their every-day teaching repertoire more in line with new ideas on competence-based education.\nResearch limitations/implications: - Only qualitative data are gathered, in one school, representing one school type, in one country only, limiting the statistical generalizability of the results. Not all respondents are able to participate in all four data sources.\nPractical implications: - Schools should develop HR policies that offer teachers CPD activities in the crucial area of competence-based education; the frameworks presented in the study can be used to this end, by teacher educators as well as by HRD professionals interested in combining training programs with informal employee learning.\nOriginality/value: - The study draws on literature from various disciplines (especially educational psychology and HRD), which traditionally have remained mostly separate. It combines insights from four separate data sources.\n",
"role": "assistant"
}
] |
Operating without operations: how is technology changing the role of the firm?
|
[
"Technology",
"Operations management",
"Service operations",
"Operations strategy",
"Service systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: A visit to Disneyland Tokyo in 2050...\nLooking back at service operations research: The literature on service operations is vast, with dedicated literature reviews including detailed classifications of service operations research (Machuca et al., 2007; Smith et al., 2007), historical accounts of the emergence and development of key ideas (Chase and Apte, 2007; Heineke and Davis, 2007; Victorino et al., 2018), as well as more critical reviews that identified gaps as a basis for outlining research agendas (Bretthauer, 2004; Chase, 1996; Hill et al., 2002; Johnston, 1999; Roth and Menor, 2003; Field et al., 2018).\nKey themes in service operations research: In a recent review in the Journal of Service Management, Victorino et al. (2018) noted that reviews ought to adopt a more interdisciplinary approach to understanding service operations from related fields such as marketing, human resources, and information systems, reflecting a more service science perspective (Spohrer and Maglio, 2008). In what follows, themes are identified that last over time and span disciplines, aiming to identify the meta-level phenomena that the field of service operations has been built on, and will continue to be built on, as well as the related assumptions that underpin them. Identifying these phenomena and corresponding assumptions is critical to understand the future of what service operations may look like. There is irrefutable evidence that changes in phenomena often challenge prior assumptions, which in turn prompt disruption and give rise to the emergence of unforeseen trends of changes. For example, new technologies are changing market structures and challenging prior assumptions about power distribution in service systems (Fehrer et al., 2018), enabling new business models that challenge the role and importance of stakeholders in service systems (Andreassen et al., 2018), and more fundamentally may represent an inflection point for the service sector similar to the industrial revolution in manufacturing (Wirtz et al., 2018). Given the interdisciplinary nature of service operations, it is also important to highlight that marketing, human resources, and information systems, as well as service science, all draw upon theoretical knowledge outside of the business and management discipline, such as psychology, sociology, engineering, as well as science, technology, and society studies (Maglio and Breidbach, 2014). The paper begins by identifying the key roles and relationships of people and technology in service operations, incorporating examples from Field et al. (2018), which also explored the future of service operations research.\nLooking forward: future of service operations: Service operations have changed dramatically over the past 100 years. Whereas firms once controlled operations largely internally, with well-defined relationships to outside suppliers and other stakeholders, modern service providers orchestrate the actions of varied stakeholders both inside and outside the firm (Normann and Ramirez, 1993). No longer is it possible to consider service in terms of simple dyadic interactions and relationships - service occurs in a complex system or ecosystem of multiple stakeholders (Maglio et al., 2009; Vargo and Lusch, 2011). No longer are traditional notions of ownership or firm-based value creation relevant as resources are routinely shared among individuals and firms, creating the need for flexible or elastic relationships (Dustdar et al., 2011). The boundaries between providers and receivers (rather than firms and customers) are blurring, and the boundaries between people and technologies are blurring (Moldovan et al., 2017). In this section, three emerging themes from the literature that are likely to inform the future of service operations research are described: complexity, orchestration, and elasticity.\nHow is technology changing the role of the firm in service operations?: From the discussion so far, it is clear that the key to the past and the future of service operations is technology. In the past, there were well-defined roles for people and technology and firms mainly managed and controlled technologies available for use in service operations. Now, this is all changing. In the future, service operations will need to respond to greater levels of variability, heterogeneity, changing and diffuse roles of actors, and the need to better integrate humans and technology. Future service systems, will, accordingly, need to facilitate the flexible or elastic orchestration of people and technologies in complex arrangements for value creation. Consider how recent technologies help people make decisions, take action, and assess value. It is not a big leap to imagine a future in which technology assumes even greater prominence, blurring the role of the firm in service operations.\nDiscussion: The three themes of complexity, orchestration, and elasticity emerged from the review of service operations literature, and observations of the changing role of technology in service operations. Consider service system complexity. Unpredictability and uncertainty of individual customers may be less important than unpredictability of entire service systems. Technologies in the service system may not be entirely (or even mainly) within the control of individual firms, but rather may represent characteristics of constantly evolving service systems. Complexity is inextricably linked to the relationship between technology and people, and the almost unfathomable pace of change suggests that isolating individual interactions between technology and people will be increasingly more difficult, and that assemblages of multiple actors may be increasingly important.\nConclusion: More conventional reviews of service operations research, encompassing the field in breadth and depth, provide invaluable resources for scholars to trace the historical development of the discipline as well as provide detailed frameworks for subsequent service operations research (e.g. Roth and Menor, 2003; Chase and Apte, 2007; Victorino et al., 2018). By contrast, the approach taken here has been to identify broad themes that tie together service operations research, question whether assumptions inherent within them will remain, and provide evidence for already emerging trends in service and service systems. Some of the ideas, issues, and questions articulated may well not come to pass. But accurately predicting the future or providing irresistibly familiar classifications, frameworks, or research agendas was not the motivation.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the history and future of service operations, with the goal to identify key theoretical and technological advances, as well as fundamental themes that can help to imagine the future of service operations in 2050.\nDesign/methodology/approach: A review of the service operations literature was undertaken to inform a discussion regarding the role that technology will play in the future of service operations.\nFindings: The future of service operations is framed in terms of three key themes - complexity, orchestration, and elasticity. The paper makes three contributions to the service science literature by: reviewing key themes underpinning extant service operations research to frame future trajectories of service operations research; elaborating a vision of service operations in 2050 based on history and technology; and outlining a research agenda for future service operations.\nPractical implications: The case of service automation is used to provide an illustration of how the three themes converge to define future service operations, and in particular, to show how technology is recasting the role of the firm.\nOriginality/value: Service operations in the next 30 years will be very different from what it was in the past 30 years. This paper differs from other review papers by identifying three key themes that will characterize and instill new insights into the future of service operations research.\n",
"role": "assistant"
}
] |
Does individual heterogeneity shape retail investor behaviour?
|
[
"Behavioural finance",
"Financial markets",
"Demography",
"Investor behaviour",
"Personal finance",
"Behavioural economics",
"D14",
"E01",
"G02",
"G11"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: In the finance literature, it is oft-discussed that individual investors exhibit heterogeneity in financial behaviour with respect to investment decisions and asset allocation. For better understanding of retail investor behaviour, it is essential to explore into the fundamental issues such as what determines retail and small investor behaviour in the financial markets. Is the behaviour of retail investors shaped by the social environment and contextual learning over the period, such as learning and training about the investment decisions through education and peers? Are their behaviours determined by more of demographic characteristics such as gender, age, and occupation, or financial well-being such as income and investment levels? This paper attempts to explore these issues in an empirical setup and contribute to the literature on behavioural finance by way of explaining the role of heterogeneity in determining retail and small investor behaviour.\nReview of literature: The issue of determinants of investor behaviour in stock market is important for several reasons. First, it is related to an emerging area of academic research popularly known as behavioural finance and economics. This domain holds equal significance for practitioners in the financial markets and personal finance industry. Second, an understanding of investor behaviour explains further the cross-sectional mechanism of asset allocation puzzle and hence demystifies asset pricing from investors' perspective. This kind of explanation is certainly deemed to help taking the argument of behavioural asset pricing approach further in terms of methodological contributions. Finally, abundant empirical research on investor behaviour has been carried out across various markets and economies, mostly using financial market data. Very little efforts have been devoted to understand investor behaviour through primary investigation approach such as survey, psychometric, and experimental methods. Such an attempt adds to an extensive body of literature in behavioural finance and economics by providing empirical evidences through these unique approaches typically considered superior for understanding behavioural issues.\nMethodology and approach: To explore whether the effect of domain-specific determinants of investor behaviour differs across the demographic and financial profile of sample retail investors, we use simple univariate statistical tests, such as paired t-test which typically compares two samples in cases where each observation in one sample has a natural partner observation in the other sample. We employ this test to check at the difference between paired values of two (or more) samples (classified by domain and demographic profiles in our case), considering the variation of values within each sample, to produce t-statistic.\nResults and discussion: Sample characteristics\nConclusions: In this paper, we studied how trading behaviour of retail investors evolves depending on their demographic and financial characteristics. Using structured survey data, we measured investor trading propensity in three domains, namely changes in fundamental factors, unexpected market events, and personal factors, and showed that they are differenced across several demographic and financial parameters. Specifically, we empirically showed that: demographic and financial factors such as gender, age, educational status, income, and investment levels determine their trading behaviour across three domains; however, features such as marital status and occupation do not play any significant role in shaping their trading behaviour; retail investors' self-perceived confidence is a function of both expected and unexpected changes in the market and personal factors, and largely determines trading behaviour of retail investors, and self-perceived confidence level and self-reported portfolio size are positively associated implying that (over-)confident retail investors tend to believe that their investment skills being superior is bound to perform better and thus they typically hold larger than average investment portfolios. That self-perceived confidence predicts investment behaviour of retail investors, supports the proposition of importance and requirement of investor education and assessment for sound investment decisions.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to survey retail investors to study the determinants of their investment behaviour and show that individual heterogeneity and financial factors such as gender, age, educational status, income, and investment levels determine their trading behaviour across three domains; however, features such as marital status and occupation do not play any significant role in shaping their trading behaviour.\nDesign/methodology/approach: Structured surveys are conducted on retail and small investors using the brokerage services of a firm. Data collected from primary methods are used for statistical analysis in ANOVA and multiple regression frameworks.\nFindings: The authors also report that retail investors' self-perceived confidence as a function of both expected and unexpected changes in the market and personal factors largely determines trading behaviour of retail investors and that self-perceived confidence level and self-reported portfolio size are positively associated implying that (over-)confident retail investors tend to believe that their investment skills being superior are bound to perform better and thus they typically hold larger than average investment portfolios.\nPractical implications: These findings are significant because research on cross-sectional variance of individual investment behaviour explains how investor heterogeneity plays a critical role in investment and asset allocation decisions. Investors, researchers, and practitioners would use the results for financial decision making specifically related to personal finance, behavioural portfolio management, and investment advisory.\nOriginality/value: This paper is an empirical approach to explore the retail investor behaviour using psychometric approach with respect to self-perceived confidence and other perceived measures of investor behaviour. The authors contribute to the emerging set of literature on investor behaviour and behavioural finance.\n",
"role": "assistant"
}
] |
Purchasing management and the role of uncertainty
|
[
"Uncertainty",
"Interdependence",
"Supply management",
"Purchasing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Uncertainty has always been a salient feature of purchasing and supply management. From a general management perspective, Cyert and March (1963, p. 119), in their behavioral theory of the firm, claimed that firms are faced with uncertainty \"regarding deliveries of suppliers.\" In relation to purchasing management, Webster and Wind (1972) pointed out the significance of uncertainty in one of the first conceptualizations of organizational buying behavior. The authors identified three dimensions of this uncertainty: uncertainty about available supply alternatives, uncertainty regarding the consequences of these alternatives and uncertainty related to the ways other firms will react. For these reasons, Farmer (1972) argued that the decisions taken by executives responsible for purchasing and supply do involve considerable risk and uncertainty.\n2. The starting point of the paper: According to many authors, the general conditions for purchasing and supply management changed considerably from the mid-1970s. At that time, the relationships to suppliers became increasingly important owing to specialization and outsourcing (e.g. Carlisle and Parker, 1989; Davis, 1993; Lamming, 1993; Lewis, 1995; Gadde and Hakansson, 2001; Araujo et al., 2016). Van Weele and Rozemeijer (1996) even described this reorientation as a \"revolution in purchasing.\" Farmer (1997, pp. 4-5) concluded that buying firms were involved in \"proactive purchasing\" implying that they sought to influence the manner in which suppliers operated, through closer collaboration. The buyer behaviors required in managing these relationships \"were far away from the hard-nosed buyer\" featuring in the era before the reorientation. Farmer (1972, p. 12) characterized the purchasing approach in that earlier period as emphasizing \"short run reaction to short run feedback.\"\n3. Core features of purchasing management, uncertainty and dependence: In this section, we introduce and define the notions of purchasing management, uncertainty and dependence: three of the central concepts in this paper. Interdependence, power and control will be presented later, since we relate these aspects primarily to the changing perceptions of dependence.\n4. Purchasing, uncertainty, and dependence before the reorientation: Before the 1970s, the literature on purchasing focused primarily on techniques and procedures with the main mission to enhance transaction efficiency. These improvements were assumed to be attained through well-developed purchasing routines (e.g. Westing et al., 1969). Among these practices can be mentioned: principles for making enquiries, how to handle tenders, calculation of economic order quantities, inventory management, contractual arrangements and supplier selection.\n5. The situation after the reorientation: 5.1 Purchasing management\n6. The impact on interdependence, power and control: In this section, we relate the discussion on uncertainty and dependence to three other inter-organizational concepts. First, previous recommendations for avoiding dependence regarded one-sided dependence on suppliers. In current organizational arrangements, close buyer-supplier relations feature interdependence, because both parties have invested in a high-involvement relationship (Bensaou, 1999; Gadde and Hakansson, 2001). These conditions also affected the perception and utilization of power. Since the power of one party is derived from the dependence of the other party, the potential power of both parties is substantial. However, owing to their interdependence, firms cannot exploit power in the same way as before, which is the second point brought up for discussion. Third, the interdependence featuring in buyer-supplier relationships affects the perspective on control. While direct ownership control of resources used to be the main option, the evolution of purchasing management made access to resources more important, implying increased dependence, and highlighting forms of indirect control.\n7. Discussion: Previous sections showed the significant interconnections between purchasing management and uncertainty. The evolution of purchasing management shifted the attention to other dimensions of uncertainty than the ones most visible before the reorientation of purchasing and supply management. In particular, the development from low involvement relationships with suppliers toward close relationships impacted on what features of uncertainty are the most salient. However, since most firms are faced with both types of supplier relationships, if not the full spectrum from mutual independence to mutual dependence (Bensaou, 1999), handling the multiple consequences of uncertainty is a complex challenge. In this section, we illustrate this challenge by discussing the role of uncertainty in relation to the purchasing strategy of companies. The dimensions of a firm's purchasing strategy can be expressed in various ways. Here, we rely on Gadde and Hakansson (2001) who distinguished three aspects: the decision whether to make or to buy, the nature of relationships with individual suppliers and the features and the organizing of the supplier base.\n8. Conclusions and implications: 8.1 Contributions\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyze the role of uncertainty in purchasing and supply management, and the changes of this role over time.\nDesign/methodology/approach: The paper is based on a literature review of the development of purchasing and supply management over time and how these issues have been related to uncertainty and dependence. This examination also required analysis of the impact of other concepts from behavioral sciences: interdependence, power and control.\nFindings: The paper shows that the relationship between purchasing management and uncertainty has changed substantially over time. Traditionally, uncertainty was avoided, while firms today are engaged in efforts of handling the consequences of uncertainty. This modification affected the features of buyer-supplier relationships, as well as the perspectives and the exploitation of power, control and dependence.\nOriginality/value: The paper demonstrates both positive and negative consequences of uncertainty, depending on the approach applied in purchasing. Moreover, the analysis shows that uncertainty cannot be avoided. Modifications of purchasing management will reduce certain types of uncertainty. But the same modification also results in increases of other forms of uncertainty.\n",
"role": "assistant"
}
] |
Improving international attractiveness of higher education institutions based on text mining and sentiment analysis
|
[
"Higher education",
"Sentiment analysis",
"Text mining",
"International student mobility",
"Topic modelling"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Throughout the last decades the promotion of international student mobility (ISM) has been one of the top priorities of the European Commission. ISM strongly benefits both academic institutions and host countries (The European Commission, 2013; Beine et al., 2014). Students who go through an international experience often remain in the host country after their studies, which contributes to the escalation of highly qualified workers who boost the country's growth (Gonzalez et al., 2011; Beine et al., 2014). Furthermore, providing studies to foreign students is a powerful way to communicate the host countries' cultural and political values (Gonzalez et al., 2011; Beine et al., 2014). However, Middle East, Asia and Latin America's competition in this field is becoming increasingly more aggressive which leads to a rapid modification of ISM's flows (The European Commission, 2013). To succeed in this competitive environment, Europe should increase higher education institutions (HEIs) attractiveness even further by effectively promoting ISM. To encourage ISM to their country, HEI should acquire a deep knowledge about students' behaviour abroad, exploring the drivers that influence students' decision making when it comes to choosing their school (Cubillo et al., 2006; Choudaha and Chang, 2012).\n2. Background and theoretical framework: Maringe and Carter (2007, p. 463) define the process of ISM decision making as a \"multistage and complex process undertaken consciously and sometimes subconsciously by a student intending to enter higher education and by which the problem of choosing a study destination and program is resolved\".\n3. SA on education and TM process: In recent years, opinions posted online have been studied using SA, a TM set of methods to uncover the valence of written text (Mostafa, 2013). TM focusses on automated knowledge discovery from non-structured or semi-structured data collections (Hearst, 1999; Sanchez et al., 2008). Written text is often filled with spelling mistakes, abbreviations and implied meanings. Although humans have the ability to overcome such obstacles, the amount of information online today is so large that it becomes a daunting task to analyse it efficiently without some sort of an automatic pattern discovery technique (Lee et al., 2010; Mostafa, 2013). TM uses natural language processing (NLP) techniques to deal with such challenges and focusses on detecting data patterns automatically (Mostafa, 2013), while SA uses such patterns to identify sentiment polarity in text, for instance, if sentiments are positive, negative or neutral (Mostafa, 2013).\n4. Method: 4.1 Data selection and extraction\n5. Results: 5.1 Hypothesis testing\n6. Conclusion: IS often rely on online sources to select a HEI abroad since they are unable to visit the school and because they find ex-students' opinions to be trustworthy (Gomes and Murphy, 2003). These sources of information may strongly influence the reputation of a HEI and consequently the decision making of students. Thus, it becomes essential for an institution to be aware of students' opinions on social media so that they can manage their reputation and meet IS needs and wants.\n",
"role": "user"
},
{
"content": "Purpose: The increasing competition among higher education institutions (HEI) has led students to conduct a more in-depth analysis to choose where to study abroad. Since students are usually unable to visit each HEIs before making their decision, they are strongly influenced by what is written by former international students (IS) on the internet. HEIs also benefit from such information online. The purpose of this paper is to provide an understanding of the drivers of HEIs success online.\nDesign/methodology/approach: Due to the increasing amount of information published online, HEIs have to use automatic techniques to search for patterns instead of analysing such information manually. The present paper uses text mining (TM) and sentiment analysis (SA) to study online reviews of IS about their HEIs. The paper studied 1938 reviews from 65 different business schools with Association to Advance Collegiate Schools of Business accreditation.\nFindings: Results show that HEIs may become more attractive online if they financially support students cost of living, provide courses in English, and promote an international environment.\nResearch limitations/implications: Despite the use of a major platform with a broad number of reviews from students around the world, other sources focussed on other types of HEIs may have been used to reinforce the findings in the current paper.\nOriginality/value: The study pioneers the use of TM and SA to highlight topics and sentiments mentioned in online reviews by students attending HEIs, clarifying how such opinions are correlated with satisfaction. Using such information, HEIs' managers may focus their efforts on promoting international attractiveness of their institutions.\n",
"role": "assistant"
}
] |
Perceived customer participation and work engagement: the path through emotional labor
|
[
"Customer orientation",
"Work engagement",
"Emotional labour",
"Perceived customer participation",
"Perceived fit with customer"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Most organizations have implicit or explicit requirements concerning which emotions employees express and how and when they express them. These requirements are seen as more central in jobs that entail high levels of interaction with customers, such as customer-service roles. When people regulate or manage their emotions in exchange for a wage, they are said to be undertaking emotional labor (Heffernan et al., 2008; Totterdell and Holman, 2003). Emotional labor can be defined as the management of feeling to create a publicly observable facial and bodily display for a wage (Hochschild, 1983).\nBack ground theory: Emotional regulation model\nModel and hypotheses: Customer participation and emotional labor\nMethod and analyses: Measures\nDiscussion: General results discussion\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine how service employee's perceived customer participation influences beneficial deep acting behaviors among service personnel while dampening the development of the less-beneficial surface acting, which subsequently affects service employees' work engagement. Specifically, the current research explores how the personal factors of customer orientation (CO) and employee-customer fit interact with the customer participation to ultimately develop either surface - or deep-acting behaviors.\nDesign/methodology/approach: Data for the study were collected from a cross-sectional sample of retail bank and insurance companies in South Korea. Questionnaires were distributed to 750 frontline employees of several banks and insurance companies. Of these, 518 questionnaires were used for further analysis.\nFindings: The results indicate that perceived customer participation exhibits the predicted negative influence on surface acting as well as the positive effect on deep acting. Using the hierarchical moderated regression approach, the interaction effect of customer participation and CO on the service employees' surface acting was found. Finally, the positive moderating effect of employee's perceived fit with customers in the relationship between perceived customer participation and employees' deep acting was supported.\nResearch limitations/implications: First, the specific service sectors chosen for this study are retail banking and insurance. Furthermore, the study was conducted among the frontline employees of banks and insurance companies in South Korea. Second, the study used single-source data, which are prone to common method variance. While the survey instrument was structured carefully with this in mind and the results suggested that method bias may not have been an issue in this study, this problem can best be avoided by collecting data from multiple sources. Third, this study is limited by its cross-sectional approach. The cross-sectional nature of the present study does not allow causal inferences.\nPractical implications: This study provides a practical implication for managers to understand the importance of customer participation for relieving the negative effects of employee emotional labor. From a practitioner standpoint, examining the relationship between customer participation and emotional labor is of great importance given the benefits and costs associated with managing customer participation. Thus, managers should magnify the positive effect of perceived customer participation on emotional labor by increasing frontline employees' understanding of customer participation. Second, the finding that CO plays a more critical role in the reduction of surface acting has important managerial implications. The recruitment and selection of frontline employees should incorporate an assessment of the level of CO. The results of this study strongly suggest that service organizations can greatly benefit from hiring individuals with a higher CO for frontline positions because CO signals a better job-person fit.\nOriginality/value: The present study is the first to link employees' perceived customer participation with their attempts at emotional labor at work and to study how those attempts lead to work engagement. This research also shows that understanding how service employees' CO moderates the effects of customer participation on beneficial deep acting and on destructive surface acting is important in that emotional labor is a potential driver of customers' emotional states and subsequent assessments of service interaction. A third contribution of this study is the use of a new concept, which will be called person-customer fit (PCF), to reflect employees' perceived fit with customers. Although many researchers have investigated the relationship between customers and frontline employees, the research has primarily focussed on the employees' perceived fit with their organization and members of the organization and overlooked the importance of PCF.\n",
"role": "assistant"
}
] |
Completely online group formation and development: small groups as socio-technical systems
|
[
"Sociotechnical change",
"Communities",
"Social capital",
"Human capital",
"Team working"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Small groups are powerful engines for getting work done. When small work groups are formed, members must get to know each other, make sense of their tasks and coordinate task progress and completion. Completely online small groups may never meet face to face, And this means they face at least two unique challenges when they are starting out. First, no technology available today recreates the experience of a shared, physical work space (Heath and Luff, 1992; Teasley et al., 2002; Birnholtz et al., 2007). Second, when completely online groups form, members must make sense of available technology, each other and the work at hand simultaneously (Laffey et al., 2006). In this study we describe how one small group meets this challenge in the context of a completely online graduate level software design course. We then compare the evolution of interaction patterns for that group with other groups assigned the same tasks and using the same tools. Our comparisons make group development in a completely online context visible, and the deep case study informs our analysis of the common patterns that emerge from the data.By describing patterns of completely online group formation and development we illustrate the smallest and arguably most productive social unit in a new form of organization; those built through distributed and networked connections instead of physical co-location. Such virtual group behavior is hard to anticipate or identify in many prominent settings like Facebook, but occurs with some regularity in online graduate level courses, where the diversity of member experience and background is similar to what we find in virtual work groups. Prior studies of virtual work group formation describe patterns of high early expectations, followed by the cyclical movement between individual and group identity oriented behaviors, similar to studies of physically co-located groups (Smith, 2005).While studies of online learning groups offer a window into patterns of group formation, sustained evolution of group work practices in virtual organizations are studied in a range of traditional organizational contexts and newer, more technologically mediated organizational forms. Watson-Manheim et al. (2002) explore the nature of virtual work as an array of continuities and discontinuities across time, space, organization and culture, referencing online learning groups as one example. Lamb et al. (2003) conclude that external contingencies are a prime driver of ICT uptake and use within firms, further noting that, in the absence of such external contingencies, managers tend to gather only data that supports their decisions. Elliott and Scacchi (2008) present an extensive virtual ethnography of free and open source software as a social movement whose goal is no less than the institutionalization of a new social order through computerization. Socio-technical examinations of Wikipedia consistently highlight the disciplined practices that evolve in this virtual space (Kittur, Chi, Pendelton, Suh and Mytkowicz, 2007; Kittur, Suh, Pendleton and Chi, 2007; Priedhorsky et al., 2007; Kittur and Kraut, 2008). Each of these examples explicates depth and variance in the emergence of organizational practices in specific socio-technical contexts.In spite of this extensive study, organizations of all kinds maintain a strong preference for collocation over virtualization (Kiesler et al., 2005). Kling and Courtright (2003, 2004) note the rarity of community formation in completely online settings. Powell et al. (2004) describe 47 studies examining virtual work teams, which individually and collectively do not build a clear picture of how virtual organizations might work most effectively. Group formation online has been predominantly studied in the aforementioned contexts like Wikipedia or FOSS, where there is a task orientation, or in online social networks, where there is a social-emotional orientation (Adamic and Adar, 2005; Backstrom et al., 2008; Lazer et al., 2009). This research is exciting because it explains new forms of socio-technical organization; but it does not address the more common situation where new technologies influence existing modes of organization.We work to close this gap by describing completely online group formation where such phenomena presently exist, while drawing on literature from communities of practice (Wenger, 1998) and small group development (McGrath, 1984; Mcgrath, 1991; McGrath and Hollingshead, 1993; Arrow et al., 2000; McGrath et al., 2000), which describe how organizational and group practices emerge and are reified. We present a single, revelatory case study analysis of one group, showing how they simultaneously make sense of their work, each other and collaborative technology (Yin, 2009). We then analyze the unique, bi-directional interaction logs produced by the collaborative technology the groups work in (explained in section 3) using time-sliced social network analysis. The changes in network structure within and between groups in this course reinforce our findings from the case study, and reveal that patterns of completely online group formation and development can be made visible.1.1 Purpose of study: completely online groups as socio-technical systems on two levels\n2. Theoretical framework: Our work is framed in theories that help us to be clear about often confused units of analysis like \"group\", \"community\" and \"network\". Our central theoretical waypoints in this regard are Brown and Duguid (1991), Wenger (1998) and Rohde et al. (2004). In Table I the high level perspectives of each theory for each type of social aggregation are shown. In the remainder of this section we elaborate on each theory, and related theories to provide a thorough overview of how our work was guided.2.1 Social theory of learning\n3. Study context and methods: 3.1 Setting\n4. Small group case study results: Data analyses make clear three patterns and two themes, which we discuss in this section. First, we identify macro themes of tool use and cooperation. Next, we describe the behavior and context of the community as it is forged in the crucible of online cooperative work. Finally, we identify salient patterns of engagement that distinguish community formation in a completely online environment from similar events chronicled by Wenger and others in non-virtual and semi-virtual worlds.4.1 Themes of tool use and cooperation\n5. Network analytical results: emergent structures: We described one group's formation and interaction in an online course using an in depth case study. We now turn to understand the extent to which group structure is visible through analysis of the bi-directional (read and post) trace data generate by CANS.We learned three things from this analysis. First, during the first group exercise, group formation is not consistently discernable from trace data in our groups. Second and most significantly, during the second exercise the trace data reveal clear group structure for four groups, and a consistent coordination and leadership structure for all groups. Third, once group activities complete, the trace data shows no groups remain clustered together, although our qualitative analysis of Alpha Group suggests they maintained a level of group interaction. Before we explain our results in detail, we will review our methodological approach for this section.5.1 Network analysis methods\n6. Discussion: 6.1 Summary\n",
"role": "user"
},
{
"content": "Purpose: - This paper has two purposes. First, to provide insight into the formation of completely online small groups, paying special attention to how their work practices develop, and how they form identity. Second, to pursue conceptual development of a more multi-level view of completely online group experience, which can be made visible through analysis of the unique interaction logging system used in this study.\nDesign/methodology/approach: - The authors conduct a mixed methods study that integrates interviews, grounded theory analysis, case study methods and social network analysis to build a multi-layered view of completely online group and community development.\nFindings: - Completely online group formation is explicated as a socio-technical system. The paper identifies themes of tool uptake and use, and patterns of interaction that accompany group formation and development of completely online group practices. These patterns show little respect for the boundaries of space and time. It then shows how groups who are paired together for two non-sequential activities develop a common internal structural arrangement in the second activity, and are viewable as groups in the larger course context in four of six cases.\nResearch limitations/implications: - The time bounded nature of the group and community, combined with the educational context limit the generalizability of these findings.\nPractical implications: - The study shows how completely online group development can be made visible. Managers of work teams and teachers who work with classrooms in completely online contexts need to recognize the dynamic structure and interaction practices of completely online teams.\nOriginality/value: - First, little research has been conducted on completely online group formation. Second, a conceptual understanding of how group members relate to one another and how groups interact with other groups in the same socio-technical context is not explored in prior work. Third, the paper performs this analysis including data from rich, contextualized usage logs, which enables greater insight into online group interactivity than prior research.\n",
"role": "assistant"
}
] |
Individual variables with an impact on knowledge sharing: the critical role of employees' ignorance
|
[
"Multinational companies",
"Knowledge sharing",
"Aerospace and defense industry",
"Employees’ ignorance",
"Moderating variables"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Fixed graphic\n1. Introduction: In the era of the knowledge economy, organizations which are innovative performers need to manage effectively both the knowledge that is already stored in various organizational repositories and new knowledge that is externally derived (Jantunen, 2005). Such efforts can maximize organizational performance by improving productivity and overall efficiency of operations (Nonaka and Takeuchi, 1995). As organizational effectiveness is built upon employees' knowledge, their involvement in knowledge sharing (e.g. face to face and virtual communities of practice) has become one of the most prominent strategies for organizations looking to manage their knowledge assets effectively. Examples to support the above statement include but are not limited to Dow Chemical, Shell, Schlumberger, Cap Gemini Ernst\n2. Research framework and hypotheses: 2.1 Organizational knowledge sharing\n3. Methodology: 3.1 Study context\n4. Discussion: In this study, it was found that employees' ignorance has a direct impact on knowledge-sharing processes, thus restricting employees from participating in knowledge-exchange and -transfer activities on a daily basis.\n5. Concluding remarks: This paper identifies a tentative link between ignorance and knowledge sharing and argues that managing employees' ignorance could improve their knowledge-sharing behavior. As very little of this discussion is captured by the current KM literature and no relationship has previously been identified between employees' ignorance and knowledge sharing within the aerospace and defense industry, it is claimed that the effectiveness of knowledge sharing could be greatly improved, by understanding what needs to be known and also by acknowledging the existence of unknowns. Research findings also lead us to differentiate three types of employees' ignorance, implying that dissimilar organizational ill-informed scenarios may differently affect knowledge sharing within organizations.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify individual variables with an impact on knowledge sharing and explore the under-discussed construct of employees' ignorance. This can enhance the knowledge-sharing process and facilitate the development of greater intellectual capital.\nDesign/methodology/approach: Eighty-four dependent variables affecting knowledge sharing are analyzed and classified into 11 categories. In addition, the direct effect of employees' ignorance on knowledge sharing is introduced and empirically investigated in a case study of a multinational organization operating within the aerospace and defense industry.\nFindings: The findings suggest that employees' ignorance may negatively affect their intention to share knowledge, thus leading to poor decision-making and communication in organizations. Employees' ignorance could also limit the organizational ability to repel external threats, implement innovation and manage future risks.\nOriginality/value: A classification scheme based on different categories of employees' ignorance is developed, providing tailor-made recommendations for practitioners facing different types of ill-informed organizational scenarios. Further, the need to shift the emphasis away from the management of knowledge to the management of ignorance is also an important contribution of this paper.\n",
"role": "assistant"
}
] |
Whiteness, ethnic privilege and migration: a Bourdieuian framework
|
[
"Migrant workers",
"UK",
"Bourdieu",
"Ethnic privilege",
"Post-socialist Europe",
"Postcolonial whiteness"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: This paper offers an empirical example of how the Bourdieuian conceptual trinity of field, habitus and capital can be utilised in order to explore how whiteness shapes ethnic privilege or disadvantage at work. Discussions of racism, whiteness and/or ethnic privilege are rarely tackled in organisational and management studies exploring workforce diversity (Acker, 2000; Cox Jr and Nkomo, 1990; Grimes, 2001; Nkomo, 1992; Tatli and Ozbilgin, 2012). There is still also a lack of focus on labour migration processes in diversity research (Al Ariss et al., 2012; Al Ariss, 2010; Al Ariss and Syed, 2011; Bell et al., 2010; Dietz, 2010; Hakak et al., 2010; Healy and Oikelome, 2011; Hosoda et al., 2012). Furthermore there is not much research that specifically explores the relationship between whiteness and migration within organisations (Leonard, 2010a, b). Nevertheless whiteness and ethnic segmentation at work have always been scrutinised in critical race and critical whiteness studies and are of growing interest to human geographers (Dyer et al., 2010; Hunter et al., 2010; Hunter, 2010; McDowell et al., 2007; McDowell, 2008; Wills et al., 2010).\nBourdieuian conceptual framework: I argue that Bourdieu's general theory of field, habitus and capital offers a useful theoretical and methodological framework for exploring the complexity of ethnic privilege and whiteness at work. Although Bourdieu has mostly been received as a theorist of class, his concepts have also been effectively used by scholars exploring racism and whiteness (Hage, 1998; Paynter, 2001; Puwar, 2004; Weiss, 2010). To move away from purely economic logic and to account for the structure and functioning of the social world, Bourdieu (1986) developed different forms of capital, namely economic, cultural, social and symbolic[2]. A Bourdieuian conception of different forms of capital provides a relational and multilevel framework for understanding capital accumulation and deployment (Al Ariss and Syed, 2011). It also provides an emic approach to intersectionality (Tatli and Ozbilgin, 2012). As such it enables simultaneous exploration of cultural and economic processes of group formation from multiple levels and standpoints.\nContextualising A8 labour migration through an exploration of the relationship amongst transnational fields: More than 20 years after the fall of the Berlin wall there is still a tendency to construct all post-socialist CEE countries and their nationals as the same[6]. This points to the continuation of colonising practices in contemporary Europe that need to be taken into account when looking at very recent history and its effects on migrant labour. In this regard Borocz (2001) argues that Eastern Enlargement was characterised by institutional elements of colonial imperial structure that combined four mechanisms of control, namely, unequal exchange, coloniality, export of governmentality and geopolitics. From this perspective accession countries were evaluated through various means and as such reinvented with the selective aggregation of facts and fiction. These have drawn on the familiar ideological Cold War divide and on the selective assessment of applicant's economic and political development with regard to an idealised EU[7] state (Kovacs, 2001; Kovacs and Kabachnik, 2001; Sher, 2001). This situated CEE countries as stuck in the past and legitimised the framing of the enlargement as the EU's civilising mission towards its inferior \"East\". This legitimisation enabled economic colonialisation to EU-based corporations that are today the biggest investors in CEE (Borocz, 2001). However, colonialisation did not manifest itself only in control over and ownership of economic assets, but also in political influence on (meta) governance (Bohle, 2006).\nMethod, sample and data analysis: This research draws on 35 semi-structured interviews with white Polish and Slovenian workers living and working in London and south England. This research project was born out of professional interests and my personal experiences as a migrant from the CEE to the UK. My own position, as a white migrant woman, constructed an imagined sameness between myself and interviewees, which in some ways enabled easier access to some participants and at times made interviewees more comfortable to talk about the way they are positioned and treated on the UK labour market. But on the other hand this also made it more difficult to talk about whiteness, because interviewees would often assume that I would understand their unspoken assumptions. In order to mitigate this I engaged with methodological tensions between the production of sameness and otherness and aimed to address these assumptions through in-depth questions that provided additional explanations or examples (Frankenberg, 1993; Gunaratnam, 2003). Interviews lasted between one and two hours and explored reasons and motivation for migration, migrants' experiences in accessing work and working in the UK and strategies to overcome difficulties arising from work and recruitment process. Workers were accessed through on-line forums, communities and organisations that connect Polish or Slovenian migrants in London or the UK, as well as recruitment agencies that specialise in supplying A8 workers in the UK. Amongst interviewees there were 23 women and 12 men, between 23 and 42 years old and with good English language skills. A summary of the sample demographics is presented in Table I.\nFindings: Below I present the findings which illustrate how whiteness shapes ethnic privilege or disadvantage at work. This also points to the usefulness of a Bourdieuian framework in exploring whiteness. I demonstrate that a Bourdieuian framework enables the exploration of processes of power on transnational fields that can uncover contextual, intersectional, transnational and relational workings of whiteness at work. The findings illuminate racial segmentation of the UK labour market, expose intersectional workings of whiteness that affect CEE workers' position and their agency and expose relational and transnational workings of whiteness and their effects on diverse workforce.\nDiscussion: By exploring how whiteness shapes ethnic privilege or disadvantage at work, this study contributes in two ways. First it provides empirical evidence of the complexity of ethnic privilege and disadvantage at work, and second it offers a Bourdieuian framework for the exploration of whiteness at work. As such this study contributes to the debate on theoretical and methodological expansions of research focusing on migration and diversity at work (Al Ariss, 2010; Al Ariss and Syed, 2011; Al Ariss et al., 2012; Metcalfe and Woodhams, 2012). It also offers new insights for research exploring the relationships between whiteness, migration and work (Dyer et al., 2010; Leonard, 2010a, b; McDowell et al., 2007; McDowell, 2008; Wills et al., 2010).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is twofold. First it offers an innovative conceptual framework for exploring how whiteness shapes ethnic privilege and disadvantage at work. Second it offers empirical evidence of the complexity of ethnic privilege and disadvantage explored through experiences of migrant workers from post-socialist Central and Eastern Europe (CEE) on the UK labour market.\nDesign/methodology/approach: Using a Bourdieuian conceptual framework the paper begins from the historical and macro socio-economic context of EU enlargement eastwards in order to explore whiteness and the complexity of ethnic privilege at work through semi-structured in-depth interviews with 35 Polish and Slovenian migrant workers in the UK.\nFindings: The findings highlight racial segmentation of the UK labour market, expose various shades of whiteness that affect CEE workers' position and their agency and point to relational and transnational workings of whiteness and their effects on diverse workforce.\nResearch limitations/implications: Research has implications for diversity policies within organisations and wider social implications for building solidarity amongst diverse labour. Future research could increase generalisation of findings and further illuminate the complexity of ethnic privilege.\nOriginality/value: The paper contributes to management and organisational literature by offering a Bourdieuian conceptual framework for analysing whiteness and the complexity of ethnic privilege at work. It uncovers intersectional, transnational and relational workings of whiteness that shape ethnic privilege and disadvantage at work and speak of ongoing colonising and racialising processes that are part of contemporary capitalism.\n",
"role": "assistant"
}
] |
Forecast competitor service strategy with service taxonomy and CI data
|
[
"Service taxonomy",
"Competitive intelligence",
"Business strategy",
"Service process",
"CI typology"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Employment in the service sector in many OECD countries, including the USA, exceeds 70 percent of the total working population, and the production output of services exceeds 70 percent of GDP (Organisation for Economic Cooperation and Development, 2005). Competition among service firms has also become fiercer than ever as the service sector has become more and more important. To gain competitive advantages, a firm needs to formulate its own business strategies by considering the competitive environment. Therefore, legally collecting competitors' information, conducting competitive information (CI) analysis and forecasting possible strategies to use as references in formulating its own strategies has become a compelling necessity for the service firm.This paper first develops a generalized service taxonomy model by extracting the four key dimensions from the service process which all services must go through during service delivery and consumption. The four key dimensions are:1. service provider;2. service transformation process;3. customer; and4. service encounter place.To further classify the service, each dimension is assigned with three elements that were identified through extensive literature reviews. The combination of each element coming respectively from the four dimensions forms one type of service, called a service module. There are a total of 81 (=34) service modules that can be used to characterize a service business.A service firm usually provides a variety of core services. Different core services may have different performance indicators such as revenue, profit, market share, and ROI (return on investment), etc. in a service firm. The CI manager must obtain these data in a legal and ethical way.Behind these performance indicators are resources underpinning them. The strategies of a service firm are formulated based on these available resources. Under the concept of resource-based strategy formulation (Hamel and Prahalad, 1994), this paper develops a mathematical process by using the performance indicator as the performance weight (PW) of a service module and summing together all the weighted service modules. Salient elements are finally selected with priorities by applying heuristics of a 20/80 rule and large number rule on the resulting sum of the elements. The salient elements with respective priority can then be interpreted to forecast a competitor's service business strategy.\nLiterature review: Service definitions\nIntegrative service taxonomy model: Service process\nIllustrative case: CI data\nDiscussion: Consideration of customer satisfaction\nConclusion: The objective of this research was to develop a quantitative model that is able to forecast a competitor's service business strategy with CI data. We first develop a service classification model, 3P+C service dimensions and their respective elements, by a qualitative methodology of extensive literature reviews. From the definitions of services made by previous researchers and the nature of services businesses, two qualitative propositions are formed. Then, based on these propositions, the combined service modules of different core services weighted by CI data such as profit ratio (PW) are integrated. Optimization heuristics of the 80/20 rule and the large number rule are used to extract the salient elements for the purpose of the effective use of resources. By interpreting the salient elements with priority coefficients, a competitor's service business strategy can then be forecast. The model also predicts that the competitor may allocate 20 percent of its current resources to salient elements in the hope that it can generate another 80 percent of the current profit. An illustrative case of a commercial bank is exemplified using the developed model.The paper develops a new CI approach alternative to the existing CI methods. The uniqueness of our approach is first to extract the essences of service process, i.e. 12 service elements. Then, mathematical processes with weights and heuristics are used to find the salient elements, i.e. the most important elements of integrated services and their relative importance. Numerical figures such as profit ratios are used to identify the relative importance of elements in numerical figures. It is more precise to compare than just qualitatively to say high, medium or low, and the model thus provides clearer directions and priorities in strategy making.The model can also be used to improve the firm's strategy of resource usage by benchmarking. To achieve it, the firm should:* analyze its own core services with our 12 service elements, and use its current profit ratios as the weights to generate the salient elements of the integrated services following our mathematical process;* obtain the profit ratio data of the same services of the most successful company in the same industry, and use the ratios as the weights to generate the referenced salient elements; and* compare its own salient elements with the referenced salient elements and see the differences of the weights of the elements.Based on such differences of salient elements, the firm can make the necessary resource adjustments, such as personnel, facility, and back office operation, etc., to improve its performance.Wright et al. (2002) found that CI activities were regarded as a crucial and integral part of company success, and argued that the best CI practice contained strategic attitude and strategic users. The staff of the whole firm, both management and operational, should support CI activities coordinated by the CI manager to monitor competitors and identify opportunities and threats in the industry as a continuous long-term mission. The firm has to track competitors, simulate their strengths and weakness, build scenarios, and plan counterattacks. CI activities have to link to the firm's decision-making process. The taxonomical model and the mathematical process developed in our paper can be used as an effective tool to assist in accomplishing such missions. Different weights can be simulated to quickly generate many different scenarios; this can be done easily using a simple computer program.The literature on CI is limited (Wright et al., 2002). Most previous studies were qualitative research at the conceptual level. Some researches used surveys and statistical methods to compare different CI practices among different countries (de Pelsmacker et al., 2005; Priporas et al., 2005; Wright and Calof, 2006). Few studies have used a mathematical approach for competitive strategy analysis. The contribution of this paper is to construct a mathematical process from service classification concept to forecast a competitor's service business strategy. As service output exceeds 70 percent of GDP in many countries, research focusing on service business strategy forecasting using CI data should have extraordinary significance.\n",
"role": "user"
},
{
"content": "Purpose: - The paper aims to develop a service taxonomy model and a mathematical process to forecast a competitor's service business strategy in a multiple service business context by inputting CI data such as the profits of existing core services.\nDesign/methodology/approach: - A qualitative method of literature review is adopted to build a service taxonomy model and form two propositions. Based on the multiple business process integration concept of the resource-based view, a mathematical process constituted by service modules and weights is developed. Salient components of the competitor's service business are identified to forecast the competitor's service business strategy after utilizing optimization heuristics of 80/20 and large number rules.\nFindings: - The model is able not only to forecast a competitor's service business strategy, but can also help develop the firm's own new service strategy. The resources of the firm can then be realigned to counteract the competitor's strategy.\nResearch limitations/implications: - The developed model is mainly applicable to service businesses. The collection and use of CI data must consider ethical issues, which might limit the sources of data.\nPractical implications: - To forecast a competitor's strategy correctly , good quality CI practice is necessary. Experienced people in the CI department are critical to the production of good quality forecasts.\nOriginality/value: - The contribution to CI impact studies is that the mathematical forecasting process is developed based on qualitative service taxonomical research. Key elements of the service process are identified as salient elements, which serve as the main focusing points in forecasting competitors' strategy.\n",
"role": "assistant"
}
] |
Female business faculty attrition: paths through the labyrinth
|
[
"Gender equity",
"Business schools",
"Attrition",
"Ethical leadership",
"Exclusionary climate"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: What do women perceive about their attrition from business schools? The under-representation of women in academia is well documented and researchers are beginning to explore it in depth (Benschop and Brouns, 2003; Bilimoria et al., 2008; Fletcher, 2007; Kaminski and Geisler, 2012; Kjeldal et al., 2005; Linehan et al., 2009; Maranto and Griffin, 2011; Xu, 2008b), though little research has been published about US business schools (for exceptions, see Fotaki, 2011, 2013; Weisenfeld and Robinson-Backmon, 2007). Across academic fields, these studies demonstrate that women are paid less, and promoted and tenured more slowly than their male counterparts. This may be especially problematic in managerialist organizational cultures (Fletcher, 2007).\nStructural explanations: Structural obstacles include practices such as gender discrimination in hiring, salary, support or other rewards, promotion and tenure, and/or a hostile organizational culture or climate (Xu, 2008b). Xu (2008b) found women faculty in STEM fields were dissatisfied with research support, advancement opportunities, and free expression of ideas. Her findings implicate academic cultures that inequitably distribute professional rewards. Likewise, women's career expectations in business academia may be frustrated over time. Bilimoria et al. (2008) found a leaky pipeline for female faculty in STEM fields, meaning that women do not attain higher academic positions. However, interventions to improve organizational culture help to fix the pipeline. Maranto and Griffin (2011) found that demographic dissimilarity, and lower perceptions of fairness, and gender equity were antecedents to an exclusionary \"chilly climate\" for female academics. August and Waltman's (2004) study across disciplines indicates environmental conditions, departmental climate, and demographics are important domains related to female faculty satisfaction and retention. Thus, we asked participants questions exploring positive and negative aspects of working in the business school (\"School\"), including culture, workload, social support, job satisfaction, and career opportunities.\nGender socialization explanations: Gender schema theory states that men and women hold unconscious, implicit schemas about sex differences that affect expectations and evaluations of men and women in the workplace and may explain women's under-representation in academia (Valian, 1998). Ibarra (1992) found that two network patterns tended to reinforce organizational power structures that entrench gender inequality:\nMethod: We used structured, open-ended interviews to gain in-depth understanding, in a real-word environment, of why female business faculty members leave an academic position. We selected women who left a business school in a Midwestern public university (School) with a faculty count in the mid-1930s that has experienced 68 percent female faculty turnover from 2000 to 2012. We identified seventeen former female faculty members of the School. Two declined, two could not be located, and one did not return our telephone calls. We gathered data from 12 women in 2012. At the time they left, one was 20-29, one was 30-39, seven were 40-49, two were 50-59, and one was over 60. The participants' ranks[1] at the time of leaving were as follows: four instructors, four assistant professors, three associate professors, and one full professor. Seven women had PhDs, four had MBAs, and one had a JD, Nine reported a voluntary separation and three reported an involuntary separation.\nResults[2]: The major emergent themes include gender socialization and structural barriers leading to attrition; specifically: curtailed career opportunities, including a lack of gender equity in promotion and tenure; an exclusionary, managerialist culture which marginalized and demoralized women; policies and practices of leadership as a key source of problems, evincing a need for ethical leadership; and the loss of a critical mass of women.\nDiscussion: Viewed through the ASA framework (Schneider et al., 1995), the women clearly left due to a lack of fit with the culture, leadership policies and practices, and lack of career opportunities. They were attracted to faculty and administrative positions based upon fit with those positions. In several cases women who did well under a prior administration could not survive a leadership and cultural shift toward managerialism. This also stopped career advancement opportunities. This may be due to a lack of social capital that is more difficult for women to attain in order to advance to senior management in higher education (HE) (Riordan, 2011; White et al., 2011). It could be due to the \"hegemonic masculinity\" that is involved in the assessment of candidates for senior HE positions (Grummell et al., 2009). White and Bagilhole (2013) note that where managerialism destabilized an organization (as appears to be the case here), one woman felt only a choice of resistance or exit. ASA would predict that women will continue to leave the school and potentially other business schools with a competitive, managerialist culture.\nLimitations and future research implications: All research designs trade off generalizability to the population studied, accuracy or precision in control and measurement, and realism, referring to the context in which behavior was observed (McGrath, 1982). Our qualitative study focusses on realism in a particular context. It is limited in that we only interviewed women who left which may result in a negative picture of the School. Men who left may also negatively experience this culture and leadership. It is not intended to have generalizability, rather to present a model for future research. To advance this future research, we present Figure 1, a model of relationships that may lead to female faculty retention. Consistent with Eagly and Carli (2007), we posit that having a critical mass of women is needed. Next, ethical leadership is posited to be associated with gender equity and justice, shown by Maranto and Griffin (2011) to be positively associated with an inclusive climate. Inclusive climate, ethical leadership, and a critical mass are posited to be associated with psychological safety which then mediates the relationship to positive job attitudes.\nConclusion: The principles for responsible management education, an initiative with over 500 participant institutions, has a working group on gender equality, recognizing the importance of this topic to management education. Certainly it will be critical to attain gender equality within business schools in order to teach gender equality to business students. We agree with Fotaki (2013) and Verbos and Humphries (2012) that the underlying cultures need to change to accomplish this. This research is a step in this direction.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore female business faculty perceptions about attrition from a business school to uncover factors that might assist in female faculty retention in business schools.\nDesign/methodology/approach: Using a qualitative study approach and guided by past literature, the paper systematically analyses open-ended responses to interview questions and notes emergent themes.\nFindings: The major themes that emerged as factors leading to attrition: first, an exclusionary and managerialist culture which marginalized and demoralized women; second, curtailed career opportunities, including a lack of gender equity in promotion and tenure; third, poor leadership; and fourth, break up of a critical mass of women. The factors then that might assist in female faculty retention are a critical mass of women, gender equity, inclusive, collaborative cultures, psychological safety, and ethical leadership. The career patterns of the women indicated that a labyrinth is an apt metaphor for their career paths.\nResearch limitations/implications: This research examines just one school from the perspective of women who left. It holds promise as the basis for future studies across business schools and to faculty within business schools to determine whether the emergent themes hold across schools.\nOriginality/value: This study examines women in business academe through the attraction-selection-attrition framework and by extending the labyrinth career metaphor to an academic setting. The paper also provides a conceptual model of female faculty retention.\n",
"role": "assistant"
}
] |
Organizational knowledge creation practice: Comprehensive and systematic processes for scale development
|
[
"Knowledge organizations",
"Knowledge creation",
"Knowledge management",
"Korea"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: The importance of enhancing organizations' learning and knowledge creation capability is continuously increasing. One persistent challenge is how investments in learning lead to organizational performance improvement, such as cycle-time reduction, customer value creation, and financial gain (Essens and Schreinemakers, 1997). Establishing valid and reliable measures of employees' learning-related behaviors, particularly in the acquisition, application, and sharing of value-adding knowledge, and in relating those behaviors to employee productivity, is a daunting but worthwhile task (Huber, 1991). Managing the productivity of knowledge workers is the ultimate source of organizational competitiveness for the future (Drucker, 1999).\nResearch purpose and problems: The purpose of this research was to develop and validate a systematic scale for measuring organizational knowledge creation practices, based on the SECI processes of the knowledge creation theory. Much attention has been paid to the legitimacy of organizational knowledge creation from socio-cognitive perspectives (Brown and Duguid, 2001; Nonaka and Takeuchi, 1995; Tsoukas and Mylonopoulos, 2004; Yanow, 2004), but little research has been conducted on the development of systematic and reliable instruments to measure the proposed practices of knowledge creation and conversion (Lloria and Moreno-Luzon, 2005; Schulze and Hoegl, 2006). Huber (1991) once called for more interdisciplinary and empirical studies on organizational learning and in response scholars developed questionnaires to provide managers with tools for assessing relevant concepts. For instance, the learning organization survey by Garvin et al. (2008) measures organizations' practices in three domains:1. supportive learning environments of psychological safety, openness, appreciation of differences, and time for reflection;2. concrete learning processes and practices of experimentation, information collection, education and training, analysis, and information transfer; and3. leadership.Similarly, the dimension of learning organization questionnaire captures information regarding members' practices in seven dimensions: continuous learning, dialogue and inquiry, team learning, empowerment, system connection, embedded system, and leadership (Yang et al., 2004). Scale development research can lead to the examination of psychometric properties of used items as well as meaningful interpretations regarding relationships among concepts examined, thereby playing an essential part of theory building (Gable and Wolf, 1993).According to Holt et al. (2007), a comprehensive approach is needed in scale development through the procedure of an integrative literature review, item development, panel expert review, item confirmation, questionnaire administration, item reduction, scale evaluation, and item confirmation. Although Nonaka et al. (1994) once used statistical approaches for validating the proposed four concepts of the SECI process and 38 knowledge creation behaviors, the work did not go through systematic procedures suggested by Holt et al. (2007), especially with regard to item development and confirmation.In their follow-up study, Nonaka and Konno (1998) added another supra concept called \"Ba,\" meaning physical, virtual, and mental shared spaces, consisting of four platforms: originating, interacting, cyber, and exercising, each of which corresponds to the SECI, emphasizing the role of contextual and social elements in knowledge conversion. Until the present, however, no independent research has been conducted to validate concepts and behaviors associated with the organizational knowledge creation theory. Finally, even though teams have been recognized as the main cognitive and temporal space to facilitate the SECI process, team-level or team-oriented behaviors did not receive due attention in the past.\nSignificance of the research: This systematic scale development of the SECI process would shed light on empirical testing of the content viability of the organizational knowledge creation theory and also present a practical instrument to measure organizational knowledge creation practices. The theory includes various concepts and behaviors expected of top-level leaders, mid-level managers, and employees within work groups for initiating and advancing knowledge assets; thus, items reflective of core behaviors of organizational knowledge creation can be used for leadership, managerial, and employee development purposes. The primary areas of this research include:* a comprehensive literature review to constrict essential domains and scale items;* a subject matter experts' (SME) panel review to verify the selected domains;* an empirical data collection and initial exploratory analysis to deduct scales; and* a confirmatory analysis to validate initially developed scales based on the model-fit to collected data set.Finally, procedures implemented in this study should illustrate a scientific and rigorous process for the development of an effective scale in the field.\nReview of the literature: Specification of items that representatively and parsimoniously capture the targeted domain is the first and most important phase of scale development (Gable and Wolf, 1993; Holt et al., 2007). To develop items that represent distinctive and important behaviors of organizational knowledge creation and conversion practices, a comprehensive literature review was conducted on the theoretical framework of organizational knowledge creation and four modes of knowledge conversion.Theoretical framework of organizational knowledge creation\nMethods and approach: In achieving its purpose of developing and validating comprehensive and systematic scales to measure the practices of organizational knowledge creation based on the knowledge conversion SECI theory (Nonaka, 1994; Nonaka and Konno, 1998; Nonaka and Takeuchi, 1995; Nonaka and Toyama, 2003, 2007), this research included two stages for scale development:1. initial development; and2. empirical validation (Hinkin, 1995, 2005).For the initial stage entitled domain and item development, a theoretical approach was taken using the integrative and comprehensive literature review, SME reviewing, and initial domain and scale development. During the second, the empirical stage of item reduction and validation, statistical approaches were used, which included administering the survey and conducting statistical data deduction, exploratory factor rotation, and the confirmatory factor analysis (CFA).Initial domain and item development\nData analysis and results: Data were analyzed using SAS 9.1, SPSS 16, and LISREL 8.80 software. The following sections summarize the results of descriptive statistics, EFA, and CFA. During the last phase, higher order CFA was also conducted to access the theoretical representativeness of four latent domains for the uni-dimensionality of knowledge creation SECI theory.Data screening process\nDiscussion and implications: The primary source of organizational performance improvement is an interaction-based organizational learning process (Song and Chermack, 2008). The SECI theory has been widely practiced to promote dynamic organizational learning, but empirical validation efforts in the past were rare. In this research, statistically acceptable scales were developed through rigorous and systematic procedures.As a result, 17 items were defined to measure the four domains of SECI knowledge creation practices. Although the EFA and maximizing factor loading processes did not support the four-factor structure suggested in the SECI theory, EFA results supported the close relationship between the combination and the internalization phases with models and solutions for products and processes through group or functional collaborations. Furthermore, the CFA supported the usefulness of the four domains of the SECI theory. Highlighted managerial and leadership practices slightly differ between those two phases - for example, managers need to justify various concepts using criteria, such as cost, profit margin, and customer relations during the combination phase (Nonaka and Takeuchi, 1995), while the leaders' provisions of cross collaboration and requisite redundancy are critical during the internalization phase for employees to internalize organizational knowledge. Given this slight difference, empirical validation of the four-factor model was more meaningful and sensible.Balanced inclusion of items reflecting individual and team members' practices of acquiring and sharing knowledge in organizational contexts is a major contribution of this study. Some of the sample questions for each of the proposed latent domains are as follows:* Socialization. \"I gather information from other departments\" and \" Our team collects work-related information and ideas from (in) formal relationships with other people\".* Externalization. \"I facilitate creative and constructive conversation among group members\" and \"Our team develops new ideas through constructive dialogue by using figures and diagrams\".* Combination. \"I engage in developing criteria to determine the value of new concepts\" and \"Our team conducts experiments and shares the newly developed concepts with the entire organization to evaluate the value of the concepts\".* Internalization. \"Our team combines existing and new concepts in meaningful ways\" and \"Our team-members use newly learned knowledge as the sources for the next time applications.\" Scores analyzed at the individual, group, or firm level can be utilized for comparison or development purposes to promote the collaborative knowledge creation practices in organizations.As mentioned in the previous section, the primary approach of this research was the deductive method based on the solid theoretical foundation of the SECI knowledge conversion theory. This particularly entailed the following procedures:* item development based on a comprehensive and integrative literature review;* item review with SME and modification processes;* a data screening process and basic descriptive analyses followed by reliability analysis and data deduction;* implementation of EFA for the identification of factor structure; and* using the confirmatory factor analyses - single-order and higher order CFAs - to measure statistically acceptable psychometric properties of the proposed theoretical structure of SECI knowledge conversion theory.This research could be considered the first statistically comprehensive attempt to develop scales of organizational knowledge creation practices, which consider both individual and team/organizational levels. These measurement items will shed light on not only more dynamic research on knowledge creation and knowledge-based performance improvement, but also more accurate application in the workplace to examine the status and process of organizational knowledge creation practices. Furthermore, leaders and managers could consider these currently proposed four domains of knowledge creation and developed measurement items for developing formal and informal learning interventions to promote a human interaction-based KM approach, knowledge-based organization development, and organizational performance improvement strategies.\nLimitations and further research: This study took a deductive approach to scale development. Nonaka et al.' knowledge conversion SECI theory was used as the foundation to develop and confirm initial domains and items. Although a comprehensive and integrative literature review was conducted on how the selected theory was aligned with other views, such as firms as socially complex systems (Antonacopoulou and Chiva, 2007; Brown and Duguid, 2001) and distributed knowledge systems (Tsoukas, 1996; Yanow, 2004), for further research, more diverse views should be considered to reflect more comprehensive perspectives of human knowledge-creating behaviors to strengthen the validity of the developed scales.Second, data were collected only from Korean for-profit organizations. However, depending on the sample characteristics and structure of the workplace and respondents' perceptions, research results could be varied. In order to ensure more adequate applications of the measures of organizational knowledge creation practices, sample diversification should be considered in terms of data collection from different types of industries and multiple locations, regions, or countries.Third, all data analyses were quantitative-oriented methods. All adapted data analysis strategies were adequate and well applied according to the research purpose and process; however, to capture a more dynamic and interactive phenomenon in the workplace, observation-based qualitative approaches should be added. This qualitative research approach would strengthen the validity and triangulation of emergent themes with measurement items, and furthermore, provide a different angle to confine missing concepts of the current knowledge.Fourth, factor analysis is a useful technique for validating theoretical constructs with data sets. Before confirmatory analyses are performed, exploratory analyses are essential for analyzing theoretical constructs in view of data obtained. This study found that responses from the socialization and the externalization were highly correlated, and the result of EFA indicated that distinctions of factor structure were not clear as the SECI theory claims. Other scholars proposed conceptual frameworks for knowledge flow within organizations (Song and Chermack, 2008; Yoon et al., 2009); thus, more studies of theory development and statistical validation should be conducted.Finally, although fidelity of the English version of the measure was ensured in terms of a systematic translation approach and a SME reviewing process, in order to increase the reliability and validity of the currently developed knowledge creation measures, a multi-cultural comparison analysis needs to be considered with English version-based data sets combined with the multi-trait multi-method approach.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to develop and validate a measurement tool for assessing organizational knowledge creation practices based on the socialization, externalization, combination, and internalization (SECI) processes of the knowledge creation theory.\nDesign/methodology/approach: - Data were collected from a sample of 455 knowledge workers from four Korean for-profit organizations, utilizing a systematic procedure, which includes: initial item and domain development based on a comprehensive literature review, reliability assessment and item deduction, and construct validity and psychometric property assessment.\nFindings: - Results show that 17 items related to individual and team members' practices of acquiring and sharing knowledge in organizational contexts measure the four domains of SECI knowledge creation practices.\nResearch limitations/implications: - Although samples from Korean business settings indicate a limited generalizability, this study's theory-grounded item specification and systematic procedure of scale development (i.e. descriptive statistics, reliability and inter-correlation analysis, exploratory and confirmatory factor analysis using the random split method) can be replicated in future similar studies or scale development research.\nOriginality/value: - Scores at the individual, group, or firm level can be utilized for comparison or development purposes to promote the collaborative knowledge creation practices in organizations.\n",
"role": "assistant"
}
] |
Driving forward green human resource management in India: A framework for sustainable organizational development
|
[
"Organizational culture",
"Green competitive advantage",
"Green human resource management",
"Sustainable organizational development"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Business organizations across our resource-scarce planet are under increasing pressure to adopt environmentally sustainable practices, such as recycling and saving energy, and India is no exception. In the under-researched area of green human resource management (HRM), HR professionals report that they lack a formal remit from senior leaders to drive this agenda forward. HR is motivated to act because formalized and sophisticated green HRM practices, when embedded into organizational culture, add value by creating extra commercial strength.\nSettling into green HRM practices: Green HRM maps out the proactive journey of promoting green behavior through tailoring the practices, systems, and policies of organizations in ways that benefit the environment and, in turn, society. This issue has positive significance for organizations since green HRM practices possess the power to pump out a feel-good factor into the marketplace, elevating an organization's image in the eyes of job candidates and customers. Indeed, a survey by jobsite MonsterTRAK.com revealed how ninety-two percent of candidates feel drawn to employers with green credentials.\nAn intervention framework to overcome success barriers: The six Indian manufacturers interviewed remain early in their development toward becoming sustainable organizations. Yet although they have a long way to go, their people's willingness to engage in green practices is evident. The key to success in building sustainable organizations is to embed this ethos in the culture, using many inputs for the green message as a way to integrate groups of people. A structure designed to facilitate a sustainable culture strategically combines green CSR activity, green HRM practices, and green supply chain management into a linked system - one that is harmoniously recognized across departments and sets the organization apart to receive a green competitive advantage.\nThe green road to competitive advantage: To enjoy the benefits of achieving a green organizational status, the initial priority is for HR to gain the empowerment of management support, so that HR and the leadership team consistently lead by example. The second stage is to improve the organization's green credentials by implementing these core green culture-sustaining HRM practices:\nCommentary: The review is based on \"Green human resource management: A framework for sustainable organizational development in an emerging economy\" by Mishra (2017), published in the International Journal of Organizational Analysis. This research paper concentrates on the developing green HRM practices of Indian manufacturers, and how deepening and widening these can boost competitive advantage. A framework of key interventions includes creating a harmony of wider corporate practices that all teams can recognize as belonging within the organization's cultural core.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to review the latest management developments across the globe and pinpoint practical implications from cutting-edge research and case studies.\nDesign/methodology/approach: This briefing is prepared by an independent writer who adds their own impartial comments and places the articles in context.\nFindings: This research paper concentrates on the developing green HRM practices of Indian manufacturers, and how deepening and widening these can boost competitive advantage. A framework of key interventions includes creating a harmony of wider corporate practices that all teams can recognize as belonging within the organization's cultural core.\nPractical implications: The paper provides strategic insights and practical thinking that have influenced some of the world's leading organizations.\nOriginality/value: The briefing saves busy executives and researchers hours of reading time by selecting only the very best, most pertinent information and presenting it in a condensed and easy-to-digest format.\n",
"role": "assistant"
}
] |
Farmer training to simultaneously increase productivity of soybean and rice in Indonesia
|
[
"Farmer field school",
"Farmers’ knowledge and skills",
"Java – Indonesia",
"Joint production",
"Rice and soybean",
"Curve of product transformation"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The agriculture is a resilient sector in Indonesia that supports economy of the country. McCulloch (2008) shows that agricultural employment in rural areas still dominated with 58 per cent of the non-poor and 75 per cent of the poor working in agriculture. Hill (2000) highlights that the sector constituted 24.5 per cent of gross domestic product (GDP) and employed 54.8 per cent of the total workers in 1981. As well, the sector provided approximately 50 per cent of employment and around 20 per cent of GDP in the 1990s.\n2. Literature review: Agricultural extension and farmer education programme are vital policy instruments for governments seeking to improve the productivity of agriculture. The extension programme promotes employment for young generation (Kabasa et al., 2015; Minde et al., 2015), and improve the livelihoods (Purushotham and Paani, 2016). Widely used in the FAO publications, agricultural extension is defined as a service or system which assists farm people, through educational procedures, in improving farming methods and techniques, increasing production efficiency and income, bettering their levels of living and lifting the social and educational standards of rural life (Swanson et al.,1997). Agricultural extension is considered as a particular branch of rural extension dealing with several economic and social aspects of farming (Zwane, 2012). Agricultural extension is the process of introducing farmers to knowledge, and technology innovations aimed at improving agricultural production and farmers' income and welfare (Purcell and Anderson, 1997). In developing countries, agricultural extension is considered an important public good that deserves public support (Nyambi et al., 2016). It has been recognised as an essential factor in promoting agricultural development (Anderson and Feder, 2007; Birkhaeuser et al., 1991).\n3. Theoretical framework: The theory of firm in microeconomics (Nicholson and Snyder, 2016; Pindyck and Rubinfeld, 2017; Jehle and Reny, 2011; Varian, 2014) underlies the analysis of this study. A dominant approach in the economic analysis of production technologies has been to identify technology with movements in the production function (Chambers, 1988; Antle and Capalbo, 1988). However, there is a criticism that an econometrically estimated production function will not provide unbiased estimators because of an endogenous use of inputs (Kumbhakar, 1988).\n4. Research method: The supply response of rice and soybean derived from profit maximisation is a linear form to make it easy to interpret the impact of each explanatory variable. The linear supply response function for both rice and soybean is respectively formulated as:(8)\n5. Result and discussion: The enhancement of farmers' knowledge and technology innovations is expected to improve the production of rice and soybean through the outward shifts of supply response curves, which are presented in Table II showing the estimated supply response functions for rice and soybean. There is no multicollinearity problem detected in the model. The multicollinearity problem is absent because the variance inflation factors are less than 10 for all independent variables (Sheather, 2009), and correlations among explanatory variables are lower than 0.80 (Gujarati, 2004). There is no unit root problem, meaning that the regression is not spurious (Phillips and Perron, 1988). There is also no problem related to autocorrelation of the residuals, as the Breusch-Godfrey criterion indicating no autocorrelation in the model. However, the problems of normality and heteroscedasticity exist, and then the models were estimated using generalised least square.\n6. Conclusion and policy implication: The Government of Indonesia has invested much funding to increase farmers' knowledge and technology innovations through non-formal education, which is called FFS. One of the main targets of FFS is to improve the level of production of rice and soybean to meet domestic need for both commodities. Farmers who cultivate rice and soybean simultaneously and sequentially using the same resources are expected to gain knowledge and technology innovations from the FFS. The innovations are then expected to contribute positive impacts on both rice and soybean productions. By using a micro-economic theory, this study demonstrates that improvement in knowledge and technology innovations has significantly escalated production of both rice and soybean simultaneously. Farmers have implemented the IPM and other improved agronomic technologies in both soybean and rice farming. The increases in soybean and rice productions were due to synergistic effect of improved technologies and enhanced capacity of farmers resulting from the serious and complete participation in FFS. The positive impact of FFS on the productivity of rice and soybean was robust because the assumption of joint production of soybean and rice using the same resources is verified. Other economic and technical factors as control variables strengthen the results.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the effect of enhanced knowledge and technology innovations, which were resulted from training, on the simultaneous production of rice and soybean in Java, Indonesia.\nDesign/methodology/approach: A model of product possibility frontier with two outputs produced using the same resources was employed. Based on the model, supply responses of soybean and rice were derived. Aggregate data consisting of 12 regions during the ten years of 2000-2009 were compiled from relevant agricultural institutions at the provincial level.\nFindings: Improvement in farmers' capacity has been able to increase production of soybean and rice simultaneously. Farmers' capacity increased after completion of the school. Knowledge and skill gained from the school have been applied to both rice and soybean farming. Other economic factors also affected the supply response of both commodities.\nResearch limitations/implications: Available data covered periods 2000-2009. However, the outcomes are still relevant to the current situation because food crops are the basic necessity. This study used secondary aggregate data, which might be less accurate than primary data. However, secondary data have the advantage concerning coverage and time span.\nPractical implications: The Government, in collaboration with non-government organisations and the private sectors, should continue to enhance farmers' capacity to increase the production of food crops to feed people in Indonesia, and over the world in general.\nOriginality/value: An analysis of joint production using a concept of product transformation curve can measure the impact of training on multi products.\n",
"role": "assistant"
}
] |
The effect of a sustainable label portfolio on consumer perception of ethicality and retail patronage
|
[
"Sustainability",
"Brand",
"Patronage",
"Label",
"Portfolio",
"Retailer",
"Consumer perceived ethicality",
"Sustainable labels"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Sustainable consumption has received increasing consumer interest in recent years (e.g. Luchs et al., 2010). Retailers and manufacturers are responding with integrated sustainability initiatives including the introduction of new products and brands with positive social and environmental attributes. Typically, however, consumers lack or cannot verify information about products' sustainability. To overcome consumers' difficult evaluation of credence products such as sustainable products, companies, public and non-profit organizations have developed labels that promote and verify product sustainability claims. Hartlieb and Jones (2009) explain that labelling schemes are crucial to expand the development of sustainable markets. In addition, as change agents, retailers have a key role to play in the development of labelling initiatives. However, to get retailers involved, it is important to understand how they benefit from such labelling schemes.\n2. Conceptual framework: Norberg et al. (2011, p. 369) define a collective label as \"a brand owned by an organization or an association, whose members use this label to identify themselves with a certain level of quality, particular geographical origin, or other characteristics set by the organization\". Collective labels may also identify products and not only organizations. Different organizations can use the same collective label. In contrast, a retailer-owned label is a brand owned and used by one retailer, which associates the retailer's products with certain characteristics that it has set and promoted. A major function of labels is to reduce the information asymmetry between sellers and buyers (e.g. Koos, 2011). This asymmetry exists when consumers cannot evaluate a product's attributes (or at very high costs), such as in the case of social and environmental attributes.\n3. Research design: The hypotheses were tested with a survey, through consumer-stated perceptions and behaviours.\n4. Analysis: We analyse our data with PLS path modelling (Lohmoller, 1989). We estimate our theoretical model with SmartPLS 2.0 (Ringle et al., 2005). Table I presents the descriptive statistics for the data per retailer.\n5. Discussion and implications: The results show that the portfolio of sustainable labels affects the CPE of the retailer significantly. The research extends the reciprocal effect of product brands on parent brands highlighted in previous studies to the context of labels. More specifically, the sustainability credence that a label gives to the product is also transferred to the retailer, which is perceived as more ethical. Collective sustainable labels (such as Fair Trade or organic) are more powerful than retailer-owned labels in improving CPE. Additional trust placed in collective labels over retailer-owned or higher awareness of the former may explain this difference. However, a portfolio of retailer-owned labels does also positively affect the CPE at least for some retailers. The awareness and credibility of the retailer-owned sustainable labels probably explains the differences across retailers. This finding complements Sirieix et al. (2013) who propose that consumers can react negatively to retailer-owned labels due to greenwashing suspicions (using the example of Tesco's Nurture label). Our results provide support to Hoek et al. (2013) who explain that many people use heuristics, which may render the use of general claims effective even if the issuer is not a trusted party. As a consequence, a consumer could react negatively when asked about a label, but still have a positive behaviour regarding the labelled product and the retailer distributing it.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to analyse the effect of distributing sustainable labels on the retailer's corporate brand. More specifically, the objectives are to investigate how the scope of a portfolio of sustainable labels affects the consumer perceived ethicality (CPE) of the retailer that distributes them and to understand how the perceived ethicality affects retail patronage.\nDesign/methodology/approach: In total, 230 individuals participated in a street intercept survey. Data were analysed with partial least squares structural equation modelling.\nFindings: Both the perceived scope of the portfolios of collective sustainable labels and retailer-owned sustainable labels improve the CPE of the retailer. In addition, the CPE of the retailer increases patronage. The portfolio of collective sustainable labels has more impact on the CPE of the retailer than the portfolio of retailer-owned labels, but the latter has more impact on retail patronage.\nResearch limitations/implications: In addition to limitations inherent to the methodology (e.g. survey based on stated behaviours), the model developed is simple and exploratory and does not include potential boundary conditions of the highlighted effects.\nPractical implications: Sustainable labels may not only contribute to product sales and product positioning, but also to position the retailer brand by improving the consumer perception of ethicality and indirectly increase retail patronage.\nOriginality/value: Anchored in the branding literature, this research is the first to conceptualize sustainable labels as a portfolio and measure their collective impact on the retailer's corporate brand and indirectly on patronage.\n",
"role": "assistant"
}
] |
Chinese consumer trust and preferences for organic labels from different regions: Evidence from real choice experiment
|
[
"Consumer preference",
"Consumer trust",
"Organic label",
"Mixed logit model",
"Real choice experiment"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Food risks and food safety incidents have become prevalent in China since the beginning of the twenty-first century (Yin, Gao and Wu, 2017). From an economic perspective, the market failure caused by information asymmetry is a significant cause of food safety problems (Darby and Karni, 1973). Relative to suppliers, consumers tend to have greater trust in independent third-party certification bodies (Albersmeier et al., 2010). Therefore, in a general sense, organic certification by third-parties should be considered a major tool for the effective elimination of information asymmetry and fundamental prevention of food safety risks because it provides more information to consumers (Golan et al., 2001). The development and use of organic labeling schemes has become an effective means for manufacturers to provide information about food quality to consumers. These methods have been argued to have enhanced food safety from the policy perspective (Janssen and Hamm, 2012).\n2. Literature review: Organic label is an important basis for consumers to identify organic food (Probst et al., 2012). Many studies have dealt with the topic of consumer preference for organic labels or food using diverse methods (such as contingent valuation method, choice experiment, conjoint analysis and experimental auction) (e.g. Hughner et al., 2007; Falguera et al., 2012; Liu et al., 2013; Chen et al., 2015). Choice experiment, which is based on the random utility theory, has now become a major tool for investigating consumer preferences for product attributes (Lancaster, 1966; Ortega et al., 2011; Xie et al., 2016). James et al. (2009) assessed the consumers' WTP for applesauce grown organically and/or locally, and results indicated that consumers were willing to pay a positive amount for the locally grown variant. Probst et al. (2012) revealed that consumers were willing to pay a premium of US$1.04 (per plate) if the food served contained only certified organic vegetables. Janssen and Hamm (2012) discussed consumer preference for different organic labels in six European countries and revealed that consumer WTP for different labels varied greatly. Hempel and Hamm (2016) combined a consumer survey with an in-store choice experiment to study the preferences of consumers for organic food and local food, and pointed out that organic-minded consumers (i.e. those who regard organic food production as important in the survey) had stronger preferences for organic as well as local products.\n3. Methodological background: 3.1 Selection of attributes\n4. Data: This study was conducted in Shandong Province, which is located in the eastern coastal area of China. It is the third most populated province in China. A significant difference in economic development exists between the eastern coastal regions and the central and western inland regions in Shandong Province, which represents an epitome of imbalance in terms of the economic development between Western and Eastern China. The investigation was conducted in the eastern (Qingdao and Rizhao), central (Zibo and Laiwu) and western (Dezhou and Heze) areas of the province. Given the small demand for organic food in rural regions in China, this study is based on a sample of urban consumers in the six cities of Shandong Province.\n5. Results: 5.1 Scores of consumer trust\n6. Conclusions and implications: 6.1 Conclusions\n",
"role": "user"
},
{
"content": "Purpose: The organic food sold in China can bear organic labels from different countries/regions. The purpose of this paper is to assess the trust and preferences of consumers for tomatoes carrying these different labels.\nDesign/methodology/approach: The data came from real choice experiments conducted in Shandong Province, China. A mixed logit model was used to analyze consumer willingness to pay (WTP).\nFindings: Results indicated that, among the four organic labels considered in this study, the highest WTP was expressed for organic label from the European Union, followed by Hong Kong's organic label, Japanese organic label and, lastly, by the Chinese mainland organic label. Consumer trust has a positive effect on their WTPs for the four organic labels. Providing consumers with information on organic can significantly lift their WTPs, and reduce the gaps between WTPs for different organic labels.\nOriginality/value: This research is of academic value and of value to food suppliers. International food marketers are recommended to equip their products with proper organic labels and initiate additional consumer education.\n",
"role": "assistant"
}
] |
How do new ventures attract and retain talented employees? The case of Shaadisaga
|
[
"Motivation",
"New ventures",
"Small businesses",
"Employee retention",
"Limited resources",
"Learning and skills",
"Challenging assignments"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Does employee motivation and satisfaction matter? This is a question which has been researched extensively in the context of large businesses. But when it comes to new ventures and start-ups, the entire attention is taken away by the founder entrepreneur. There is little research and guidance into the role of the employees in the functioning of new ventures.\nFive pointers to employee motivation and retention in Shaadisaga: In line with the recommendation offered by Hill and Stewart (2000) that there is a need to understand how small firms and new ventures create their own HRD best practices, we present the case of HRD best practices in Shaadisaga. We highlight the five factors which have enabled Shaadisaga to successfully retain their talented and valuable employees. We link the factors back to theory to strengthen the insights generated from the case.\nConclusion: Our findings are in line with the existing limited research on employee motivation in small companies and new ventures. In the seminal work on goal setting by Locke (2004), various mechanisms of linking incentives to goals are discussed. One such mechanism which is most relevant to most new ventures including Shaadisaga is to \"motivate by goals but pay for performance\". It is imperative for start-ups with limited resources to keep their employees motivated by giving them challenging meaningful work, but incentivize them only when they are able to accomplish the targets bestowed to them.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to make an attempt to highlight how small businesses and new ventures, which are invariably resource-constrained and cannot match the salaries of their larger counterparts, attract and retain talented employees. The study uses the setting of a new Indian venture called Shaadisaga to demonstrate successful employee development practices for new ventures.\nDesign/methodology/approach: Findings are based on in-depth interviews with the team of Shaadisaga.\nFindings: Findings reveal that new ventures - which operate with limited resources and market presence - need to leverage their human resources to their highest potential. The only way to hire and retain good talent is to keep the employees motivated, offer challenging assignments which enhance their learning and skills and make them partners of the growth story of the company.\nPractical implications: Managers and entrepreneurs who are building new ventures need to pay special attention to the development of their employees. Long-term promising careers, challenging assignments, diverse roles and a share in the growth of the company can help organizations retain good talent and get the best out of them.\nOriginality/value: Extant research offers insights into how to manage and motivate people in large organizations. However, little is still known about the management practices of new ventures, especially of those operating in the emerging economies. This study tries to highlight the human resource practices of a new venture that has been successfully building a team of motivated employees delivering to the expectations of the organization.\n",
"role": "assistant"
}
] |
The formation and effect of attitude importance in professional sport
|
[
"Sporting events",
"Leisure activities",
"Consumer behaviour",
"United States of America"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Few hedonic service experiences fuel greater interest or economic investment than professional sports do (Hightower et al., 2002; Baade, 2003). In the USA alone, 35percent of the adult population (72.1 million) regularly attend sporting events and spend approximately $14 billion on admission (US Census Bureau, 2005). For many of these fans the experience of going to a sports stadium takes on great importance (Pons et al., 2006), similar to the excitement of attending a political rally or the symbolic significance of going to a religious event (Novak, 1994). Although individuals perceive and weigh some consumptive experiences differently than others (Babin et al., 1994), studies that strictly examine how judgments of importance form from spectator experiences are decidedly absent. This neglect leaves no answer to Bloch and Richins (1983, p. 77) dated call for consumer research on leisure activity and its relationship to perceptions of importance.Many consumers attach considerable importance to their leisure experiences (Arnould and Price, 1993). Attitude strength researchers typically depict importance as a stable, resilient disposition that has a \"strong influence on social perception and behavior\" (Boninger et al., 1995, p. 61). These effects are readily observed at sporting events where short-skilled performances often stimulate strong emotional reactions in otherwise passive participants (Deighton, 1992). Early work by social psychologists weighed the degree of importance people attached to social issues like abortion or capital punishment (Krosnick, 1988a), but it is the ability of this attitude property to influence and explain key consumer activities (e.g. the processing of market communications, pre-purchase search behavior, product evaluation) that continues to attract attention from marketers (Bloch and Richins, 1983). Our paper responds to this interest by examining how attitude importance forms and effects consumers of professional sport events.\nAttitude importance: The term \"attitude\" has various definitions yet one common theme is that they entail an evaluative component or judgment (Fazio, 2007). Eagly and Chaiken (2007, p. 585) define attitude as \"a psychological tendency that is expressed by evaluating a particular entity with some degree of favour or disfavour\". Consumer researchers continue to examine how attitudes towards products and services develop because knowing how consumers' think and feel about objects enables them to predict consumer choice and strategically influence behavior (Dick and Basu, 1994; Kraus, 1995). A great deal of marketing effort aspires to create and maintain strong and enduring dispositions in customers (Bloch, 1986; Bristow and Sebastian, 2001).The properties inherent to strong attitudes have attracted a lot of empirical research. Strength theorists view importance as a key component of this focus (Krosnick et al., 1993). Summarizing 87 studies on the matter, Krosnick and Abelson (1992) characterized attitudes as important when they were salient, weighty (possessing psychological significance or value), and accessible from memory (Bizer and Krosnick, 2001). Bassili (1996, p. 638) described attitude importance as an implicit cognitive process, and classified it as a meta-attitude that forms from multiple judgments (i.e. impressions respondents hold of their own attitudes). To attach great importance fixes deep concern to matters linked to the attitude, and displays little subtlety as consumers are well aware of their avid disposition (Krosnick (1988b)). Consumer research on the nature of importance in product evaluation initially assessed it narrowly, as the weight patrons attached to the performance of individual product attributes (Jaccard et al., 1986; Neslin, 1981). Subsequent inquiries paint a broader view, having observed that consumers tend to bundle (value product attributes collectively) and sum those judgments to develop an overall meta-attitude of importance (Janiszewski and Cunha, 2004).Consumer research on importance\nAttitude activation: Several theoretical frameworks have been suggested for examining how attitude importance develops (e.g. Boninger et al., 1995). Bloch and Richins (1983) used three consecutive sectors, source-importance-response, to study the form and effect of importance. In the Source sector, perceptions originated from characteristics of the consumer and the product. Cohen and Reed (2006) identified the interplay of personal aspects and salient features of the object (product) in attitude generation. Source characteristics largely shape how consumers perceive and value certain product stimuli, and prompt importance. Once activated, the meta-attitude's cognitive process shapes and is reflected in interrelated judgments of product and brand importance. The final sector reports the effects of importance on two types of outcomes, ongoing and task-related reactions in consumers. Our study adapts this three-phase approach to examine attitude importance at sporting events (see Figure 1). In this context, spectator perceptions of a sport product's technical and functional service elements activate attitude importance (Kelley and Turley, 2001; Mason, 1999). Once triggered the meta-attitude manifests in judgments of interest in the sport and importance of a team (e.g. I am very interested in English Premier League Soccer and a strong supporter of Chelsea), thereby fueling ongoing and task-related responses in fans.Technical product\nResponses to importance: Like consumer theorists, attitude strength researchers describe importance as a consequence of perceptions that first fix the attitude's status and then activate its use when \"processing information, making decisions, and taking action\" (Boninger et al., 1995, p. 62). This pictures importance mediating between sources and responses so that reactions to perceptions of the product would first correspond to attitude change in product or brand indicators before other responses could occur (Krosnick, 1988b). Studies cite the capacity for importance to have several effects. Evidence suggests its operation enhances the acquisition of relevant information (Jacoby et al., 1978), increases message elaboration (Holbrook et al., 2005), speeds retrieval from memory (Bizer and Krosnick, 2001), improves attitude-behavior consistency and predicts behavioral response (Visser et al., 2003). Collectively, these qualities imply attitude importance acts as a type of cognitive catalyst that orchestrates product/brand judgments and directs reactions in consumers.Spectator research also ties importance to behavioral responses like ongoing media use and attendance (Burnett et al., 1993). TV viewing, reading magazines, Internet use (information search behavior) along with event attendance (purchase behavior) have also been linked to product interest (Funk et al., 2004). Task-related responses are different types of outcomes, as they entail emotional reactions to specific events. When attitude importance is strong and reflected in product/brand judgments, the scope and level of these two reactions increase. For example, a baseball enthusiast's interest in a game certainly provokes ongoing responses like attendance or media use. However, an assessment of responses would be incomplete if it failed to consider the ability of team attitude to provoke emotional task-related response (i.e. heightened levels of grief or joy to a game's outcome: Fisher and Wakefield, 1998). A sporting contest is a good example of how an important consumptive experience might evoke an emotional response (Madrigal, 2003). Depending on match outcome, importance can heighten a sense of personal achievement or a loss of self-esteem in fans (Kahle et al., 1996). Although Cialdini et al. (1976) did not directly assess importance they did describe the nature of these emotional reactions in fans with a series of experiments (e.g. \"basking in reflected glory\", Cialdini et al., 1976, p. 366). The sum of this discussion prompts the third hypothesis.H3. Attitude importance, reflected in strong positive product/brand judgments, increases ongoing and task-related responses in spectators.\nMethod: The objective of the data collection was to measure variables outlined in Figure 1. Data collection took place at a newly built sport venue in a medium sized Mid-Atlantic city in the USA The venue serves as the home for a women's professional soccer team in its second year of operation.Participants\nResults: Measurement details\nDiscussion and implications: The present study illustrates the capacity for Attitude Importance to stimulate consumer responses. Findings largely support a proposed MIMIC model and demonstrate that technical and functional attributes of the product collectively bundle to activate the meta-attitude. Once triggered the disposition's cognitive processes act by attaching significance to related judgments, which prompt ongoing and task-related responses. Although the data and the structural equations used do not establish causality, they do show \"whether the causal assumptions embedded in the model match a sample\" (Bollen, 1989, p. 4). In this sense the proposed MIMIC model fit the data more effectively than two theoretical alternatives, rendering a specific account of this meta-attitude's makeup and effect. Structural support for the source-importance-response sequence refines earlier descriptions, noting that in spectator sport attitude importance is evident when fans take interest in the sport and attach meaning to a team. The potency of these dual reflectors to prompt attendance, media use or emotional response increases when attributes of product meaning and experience become more (salient) desirable to patrons. Discussion and implications of the study's three research questions follow.Hypothesis 1: Dual reflectors\nFuture research: According to several researchers, distinguishing attitude importance and involvement is another matter that requires serious attention. Fisher and Wakefield (1998, p. 27) believe \"involvement reflects the importance of the object, situation or action\", whereas Mittal (1995, p. 666) adapted Zaichkowsky's (1985) scale to show how attitude importance toward products results from involvement (i.e. strong product beliefs). Howard and Sheth (1969) report that attitudes of importance were errantly used as a \"proxy\" for involvement. Several researchers have used importance as one of involvement's gauges (Laurent and Kapferer, 1985; Zaichkowsky, 1985). Bloch and Richins (1983, p. 73) attempt to differentiate the two, observing that importance exists as a \"cognitive state of awareness or realization that actions concerning a product matter\" whereas involvement acts as a motivational state which arouses that disposition. Our perspective differs somewhat here as it asserts that attitude importance first exists as a larger meta-attitude before guiding related judgments. However, we do concur with Mittal (1995) inasmuch as we too found the disposition arose from strong beliefs of the product.Future work on the meta-attitude's external nature could pursue other areas. First, answering questions on \"generalizability\" over time and validity in different consumptive contexts are necessary (Winer, 1999, p. 350). This could entail verifying of attitude importance with longitudinal panels or use different purchase contexts (hedonic or utilitarian consumption) to retest our framework (Babin et al., 1994). An essential next step would be to fit our supported model to other service samples to determine whether patterns hold beyond the sports arena, in less emotive utilitarian settings or on a less realistic stage (e.g. see classifications in Deighton, 1992). Second, the \"robustness\" of our findings needs attention, testing whether personal characteristics such as personality, gender or age effect the cognitive processing of attitude importance, or whether some factors predispose patrons to attach importance to judgments at one level over another. Third, the \"realism\" of our measures and methods can also be revisited. There are other more direct ways to assess importance (Jaccard et al., 1986), but this study used a MIMIC model to indirectly reflect the disposition's cognitive processing (Bassili, 1996, p. 638; Diamantopoulos and Winklhofer, 2001). Capturing the attitude strength construct in this manner commends revisiting past tests of its connection with related strength properties (Krosnick et al., 1993).\nConclusion: This study extends earlier research by demonstrating how attitude importance forms and influences responses in sport spectators. We feel a clear delineation of the construct's dual reflectors sheds further light on the interaction between consumer and context, and offers a clearer sense of how the meta-attitude develops and directs ongoing and task-related responses. Practically, discussing manifestations of importance in terms of product and brand attitude gives marketers additional insights on competitive position. Although the work presented here is preliminary and narrow in scope, our hope is that it may reopen discussion on what we feel is a fresh perspective of importance.\n",
"role": "user"
},
{
"content": "Purpose: - Current research has largely overlooked importance as a meta-attitude consumers develop from related judgments. Drawing from observations by consumer theorists and attitude strength researchers, the present study seeks to investigate the formation and effect of attitude importance in an experiential setting, spectator sport.\nDesign/methodology/approach: - The study adapts a stimulus-response framework to conduct a structural examination of attitude importance. The investigation includes a multi-stage sampling procedure that distributed surveys to spectators attending five professional sport matches (n=370).\nFindings: - Path analysis of a multiple indicator-multiple cause (MIMIC) model revealed that perceptions of technical and functional aspects of the service experience fuel a meta-attitude of importance. When evident in dual judgments of product interest and brand importance, the construct is able to play a significant role in patron responses.\nResearch limitations/implications: - These findings offer insight on the nature of importance and its role in moderating spectator behavior. Support for the structural sequence also holds implications for researchers interested in delineating other strong attitudes. However, study findings are limited to hedonic service consumers and await replication in other product settings.\nPractical implications: - Practical implications consider different mixes of dual judgments and strategies organizations might use to leverage a meta-attitude of importance in their patrons. Examples of scenario-based challenges to managing this disposition in the sport industry and in other consumer contexts are discussed.\nOriginality/value: - Despite early attention by marketing practitioners on the importance of individual product features, explanations of how a larger meta-attitude forms and affects customers are rare. The study developed a MIMIC model and used path analysis to address the matter.\n",
"role": "assistant"
}
] |
Contrasting realities: corporate environmental disclosure and stakeholder-released information
|
[
"Stakeholders",
"Environmental reporting",
"Disclosure patterns",
"Informational dynamics"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: With the variety of media outlets available nowadays, corporations can employ multiple ways to communicate with their stakeholders, aside from the traditional annual report. Through these media sources, corporations portray their environmental performance based on their own motivations (Buhr, 2007). Similarly, advances in global communications, coupled with the development of diverse social movements, have increased the availability of non-corporate information about corporations (Dey, 2007). Non-corporate disclosers provide their standpoint on corporate sustainability performance in their information releases (Gallhofer et al., 2006; Dey, 2007). Multiple views about corporate environmental performance are therefore available for stakeholders to consult and reflect upon. What kind of portrayal emerges from these multiple views? What appraisal of the corporation does this portrayal provide us? The purpose of this paper is to examine these issues by studying environmental information released by a firm and by its stakeholders, following suggestions from Adams (2004) and Dey (2007).\n2. Background and conceptual framework: Stakeholders and corporations have different approaches to environmental disclosure. Each stakeholder is concerned with specific environmental issues pertaining to its own situation (Tilt, 2007). Stakeholder disclosures reflect these particular concerns and as such are often focused, partial and partisan (Gallhofer et al., 2006; Dey, 2007). Alternatively, corporations have to manage the environmental concerns of multiple stakeholders through their operations (Freeman, 1984; Donaldson and Preston, 1995). This context is reflected in corporate reporting. Corporate environmental disclosure both influences and is influenced by stakeholders (Adams and Whelan, 2009). For example, ENGOs are important users of corporate environmental disclosure (Tilt, 1994). Many of these groups view this disclosure as a means to \"either investigate if reporting was in line with actual corporate social and environmental impacts or to facilitate pressure on companies to improve their social and environmental performance\" (O'Dwyer et al., 2005, p. 781). Their concerns and pressures influence corporate sustainability disclosure (Deegan and Blomquist, 2006). Similarly, community concerns as well as expectations from investors and governments affect firms' environmental reporting (Roberts, 1992; Brown and Deegan, 1998). This context leads corporations to disclose information covering a variety of general issues intended to respond to the heterogeneous concerns of a broad range of stakeholders (Cormier et al., 2004).\n3. Research design: Comparing corporate disclosure to stakeholder-released information about the case firm allows the analysis to move away from an organization-centered portrayal of environmental performance towards a \"polyvocal, stakeholder-driven view of (and response to) the organization's behavior\" (Dey, 2003, p. 8, emphasis added), in line with the informational dynamics of environmental reporting (Buhr, 2007; Georgakopoulos and Thomson, 2008). The stakeholder groups analysed, investors, governments, the community and ENGOs, were chosen because of their recognition in prior stakeholder literature (Donaldson and Preston, 1995), their interest in and influence on environmental reporting (Tilt, 2007), and the importance to consider them when evaluating corporate environmental performance (Ilinitch et al., 1998). Disclosure is analysed over a period of three years (2005 to 2007) in an attempt to understand the dynamics of environmental reporting. Only publicly available disclosure is considered, although private disclosure may occur between a firm and its stakeholders (Solomon and Solomon, 2006). This enables the study to concentrate on public flows of information among the perspectives and to portray public informational dynamics.\n4. The informational dynamics of environmental reporting: This section presents the analysis of the informational dynamics engaging AC and its stakeholders around environmental issues from 2005 to 2007. A background of the industry and the company is presented first to position the dynamics within the broader social context (Tregidga and Milne, 2006; Ferguson, 2007; Laine, 2009), along with a brief description of the constituents to highlight their different interests (Tilt, 2007). The dynamics are then described as a whole before being decomposed based on the timeline of disclosure. The outcome of the dynamics is then discussed, to highlight the appraisal emerging from the construction of AC's environmental performance based on multiple reports.\n5. Conclusion: The purpose of this paper was to contrast corporate and stakeholder environmental disclosures to provide some insights into the informational dynamics in which they engage. The case study comparing AC's reports to governments', the community's, ENGOs' and investors' information releases reveals that the informational dynamics in which these constituents participate are composed of multiple related patterns. In the confirmation pattern, stakeholders disclose a message corresponding to the corporate message. In the complement pattern, stakeholders provide information completing corporate disclosure. Finally, in the opposition pattern, stakeholders contradict corporate disclosure. Different patterns are associated with different levels of interactions from stakeholders, who are most involved when they combine disclosure patterns around the key issues of forest certification and logging practices. Limited interactions are observed from the case firm through all disclosure patterns, suggesting a symbolic engagement within the dynamics associated with the adoption of a strategic accountability approach.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to study the informational dynamics that take place between a firm and its stakeholders with respect to corporate environmental management.\nDesign/methodology/approach: The analysis is based on a case study contrasting environmental information reported by the case firm with environmental information about the firm disclosed by four stakeholder groups or their representatives (governments, the community, environmental non-governmental organizations and investors) over three years. The information flow of disclosure is also considered.\nFindings: The results suggest that the informational dynamics are composed of multiple related patterns. The patterns range from correspondence between disclosures to stakeholders complementing or contradicting corporate disclosures. Different patterns are associated with different levels of interactions from stakeholders, who are most involved when they combine disclosure patterns around key environmental issues for the forest industry. Limited interactions are observed from the firm, suggesting a symbolic engagement within the dynamics and a strategic accountability approach.\nResearch limitations/implications: Limitations are found in the focus on disclosure outlets without examination of their production and reception, and in the inherent nature of the documents collected to represent each perspective. Some stakeholder groups were excluded from the study due to data unavailability.\nOriginality/value: This paper offers an in-depth analysis of firm-stakeholders interactions with respect to environmental reporting and maps the information flow of their disclosure.\n",
"role": "assistant"
}
] |
Social capital and fiscal performance in India during 1991-2012
|
[
"Fiscal deficit",
"Vote turnover",
"Civil society",
"Cooperation",
"Collective action",
"Fiscal performance",
"Social capital",
"Trust",
"Principal component analysis (PCA)",
"Panel data analysis",
"H3",
"O1",
"Z"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Several provincial sub-national state governments in India accumulated large fiscal deficits or government debts during 1991-2012. Why did provincial sub-national states in India raise large debts during 1991-2012? This paper focuses on possible explanation to this question analyzing the role of social capital[1] and performance of political or fiscal institutions in India. These institutions are particularly important for making fiscal policy decisions. Fiscal institutions create the environment, the incentives, the rules and the regulations under which budgets are drafted, approved and implemented. Fiscal institutions promote fiscal discipline through proper design of fiscal policy[2]. Sometimes social and institutional framework may produce soft budget constraints that provide incentives for loose fiscal discipline. These rules highly vary over provincial sub-national state governments and thus provide a reasonable explanation for cross-section variations in debt levels or fiscal deficit performance levels[3] (Alesina and Perotti, 1995; Poterba and von Hagen, 1999; Persson and Tabellini, 2001).\n2. Literature review: The concept of social capital evolves over time[4] and it has been defined in several ways. Social capital is the network of objective relationships (structural dimension) that people and organizations have, and subjective relationships (cognitive dimension) that linked to norms and trust in other people and organizations. Initially, the social organization is considered as social capital, \"facilitating the achievement of goals that could not be achieved in its absence or could be achieved only at a higher cost\" (Coleman, 1990). Putnam (1993) provides similar characterization referring to \"the features of social organization, such as trust, norms, and networks that can improve the efficiency of society,\" whereas Fukuyama (1995) argues for certain shared norms and values that assure \"the meeting of obligation and reciprocity.\" Putnam (2000) introduces the idea of social capital in terms of \"connections among individuals - social networks and the norms of reciprocity and trustworthiness that arise from them.\" At the individual level, social capital also refers to a system of interpersonal networks (Dasgupta, 2005), which enhances cooperation and collaboration that also helps to create the economic opportunities. Social capital might be considered as \"resources embedded in social networks and accessed and used by actors for actions\" (Lin, 2001). Social capital can be considered as the stock of active connections among individuals and also helps to resolve collective problems easily in the form of \"cooperative action\" (Cohen and Prusak, 2001).\n3. Data and methodology: State-wise general election data are taken from the Election Commission of India for the period of 1991-2012. Vote turnover is measured as a percentage of votes casted in the general elections. Fiscal data are mainly revenue and expenditure of state governments in India. State fiscal data[12] are available from annual state budgets in every year. Fiscal performance of states-related data is collected mainly from Reserve Bank of India. Other variables are collected from relevant departments[13] of the Government of India. All are time series data for each state. We compile all and obtain a panel data. So, we apply panel data analysis technique which is appropriate for this study.\n4. Results: Table I describes the summary statistics of our focus variables in panel data set for the study period of 1991-2012. Fiscal deficit is the difference between expenditure and revenue measured rupees (Crore) in constant price at base year 2004-2005. Average value of fiscal deficit for the entire period is Rs43,552 but standard deviation is very high. Percentage of vote turnovers in the general election is an important variable in this study and indicates trust on political institution. Vote turnover rate varies widely (Table I) because of several reasons for different states (Feld and Matsusaka, 2003; Feld et al., 2004). Average vote turnover rates are 61.28 and 69.81 percent in Lok Sabha and Vidhan Sabha, respectively. So, more than 60 percent voters trust on political institutions (Assembly and Parliament).\n5. Conclusion: This paper provides evidence for the hypothesis that social capital influences fiscal performance. Empirical results in India suggest that social capital indirectly controls the fiscal performance of the elected government. Fiscal deficit reduces by 2.6-2.8 percent if one percent vote turnover rate rises. The strong political trust is established through vote turnover and vote share in the election for formation of government that creates the platform for sound fiscal policy decisions and hence the lower public deficits. The results are robust to a number of control variables and inclusion of additional variables. We use information from direct voter participation on political issues and government formation. In order to take both aspects of trust into account - trust among members of a society and trust between the incumbent and the constituency - we collected data from all general elections held during 1991-2012. These results are consistent with those reported by Putnam et al. (1993) from Italian regions, Knack and Keefer (1997) or Zak and Knack (2001) from cross-country regressions or Knack (1999) from US state governments for government performance.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to examine the relationship between social capital and fiscal performance using provincial sub-national state-level data in India during 1991-2012. Fiscal performance in India is based on social trust on fiscal institutions that emphasizes mainly social need for economic development.\nDesign/methodology/approach: People participation in public affairs or simply vote turnover in general election in India is taken to measure social trust on fiscal institution. Applying principal component analysis, the author constructs social capital index and examines the said relation also. Models are estimated using panel data techniques.\nFindings: Strong social capital reduces fiscal deficits. As one percent vote turnover rate rises, fiscal deficit reduces by 2.6-2.8 percent during 1991-2012. The empirical findings suggest that social capital indirectly controls the fiscal performance of the elected government. The results are robust to a number of control variables.\nOriginality/value: The strong political trust is established through high turnover rate and vote share in the election for formation of government that creates the platform for sound fiscal policy decisions.\n",
"role": "assistant"
}
] |
Segmentation of organic food buyers: an emergent market perspective
|
[
"Consumer behaviour",
"Market segmentation",
"Organic food",
"Novice organic food industry"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: The consumption of organic food has grown remarkably, both in developed and developing countries. According to Willer and Lernoud (2016), the total value of the EU organic retail market has more than doubled in the last ten years, reaching the value of 24 billion euros in 2014. The leading EU countries regarding sales were Germany (7.9 billion euros), France (4.8 billion euros), UK (2.3 billion euros), and Italy (2.1 billion euros). The highest per capita consumption of organic food was in Switzerland (221 euros), Luxembourg (164 euros), and Denmark (163 euros). The consumption of organic food is on the rise due to increasing concerns regarding the negative externalities associated with the effects of intensive farming systems on both human health and the environment (Nasir and Karakaya, 2014). Taking into account the rapid growth of the market for organically produced food, understanding a variety of motivations, perceptions, and attitudes consumers hold regarding organic food is essential for organic product producers, suppliers, marketing specialists, and policymakers. Previous research suggests that organic food purchases are mainly driven by concerns about health, the environment, food safety, and animal welfare, desires for supporting the local economy and a wholesome lifestyle, nostalgia, and pursuit of fashion (Hughner et al., 2007; Schaer et al., 2013). A frequent topic of many studies in organic food marketing is an attempt to identify characteristics of organic food consumers for segmentation purposes. Numerous studies have explored consumer characteristics associated with the acceptance of organic food including benefits sought from organic food (e.g. Lund et al., 2012), the frequency of organic food purchases, demographics, food-related lifestyles, and values and lifestyle (e.g. Nasir and Karakaya, 2014; Nikolic et al., 2014). Although the literature on organic food market segmentation is well-established in the context of Western European mature organic markets, segmentation studies of novice organic food markets of South-Eastern Europe region are difficult to find in the academic literature (Hemmerling et al., 2015). Considering the uneven stages of organic food market development in Western European countries and Southeast European countries, it is valuable to have more detailed studies of consumption within a market that is still in its infancy (Paul and Rana, 2012).\n2. Literate review: 2.1 Segmentation of organic food consumers\n3. Methodology: 3.1 Sample and data collection\n4. Research findings: 4.1 Measurement model\n5. Conclusion and implications: This study is an exploratory one that attempts to identify and describe segments of organic food buyers in a newly developing market within Southeast European region, Bosnia and Herzegovina. The present study revealed four distinct market segments of organic food buyers with a unique set of beliefs, attitudes and consumption patterns. Results indicate that, according to individual-level factors, the organic food market is comprised of four segments: \"enthusiastic social-seekers\", hostile seldom shoppers', \"hostile heavy shoppers\", and \"enthusiastic moralists\". The findings of this study have implications for research in the area of organic food marketing. Although a significant number of studies have been conducted to identify distinct segments of organic food consumer only few studies have explored the organic food market segmentation in the context of countries where organic food markets are in early stage of development. (e.g. Grubor and Djokic, 2016; Imami et al., 2017; Skreli et al., 2017). Thus, this study adds to insufficient knowledge we have related to the segmentation of emerging organic food markets in academia. Also, this study found that subjective norms play an important role in shaping attitudes towards organic food, as in the case of the \"enthusiastic social-seekers\" and \"enthusiastic moralists\". This finding is consistent with past studies which stated that subjective norms (perceived social pressure/influence) are significantly related with consumer attitudes towards organic food purchases (e.g. Al-Swidi et al., 2014; Ruiz de Maya et al., 2011). Moreover, Bosnia and Herzegovina is known as a highly collectivist society, so it not surprising to find segments of consumers who rely on the approvals of their referents in decision making regarding the trial/purchase of organic food. Also, the present study revealed that favourable/positive attitudes towards organic food purchases will not always lead to an increase in demand for organic food, as in the case of the \"enthusiastic moralists\". This finding points out the discrepancy between consumers' favourable/positive attitudes towards organic food purchases and their relatively low levels of purchase, also known as an attitude-behaviour gap (Aschemann-Witzel and Niebuhr Aagaard, 2014). Although there are a number of possible explanations why favourable/positive attitudes do not always translate into organic food purchases (e.g. the high price of organic food, unavailability of organic food, consumers' distrust on the reliability of organic food certifications and advertising claims) without further empirical work they will remain in the realm of speculation. Furthermore, findings of this study contradicts with previous research which stated that level of education and income are positively related with actual purchasing behaviour (Grubor and Djokic, 2016; Hughner et al., 2007; Krystallis et al., 2006; Yiridoe et al., 2005; Zepeda and Li, 2007; Dettmann and Dimitri, 2009). More precisely, the demographic profile of segments \"hostile heavy shoppers\" and \"enthusiastic social-seeker\" indicates that consumers with lower levels of education and income are more dedicated to organic food purchases. These results are very similar to Lea and Worsley (2005) who argued that education has minimal effect on attitudes towards organic food purchases and actual purchases. Also, these results are in line with Durham (2007) who emphasised that a high level of education has a negative influence on organic consumption. Finally, it was found that high level of social and ethical compliance is not always associated with the organic food purchasing behaviour, as in the case of \"hostile heavy shoppers\". The possible explanation for this finding is that organic food purchases are driven not only by altruistic motives (e.g. environmental concerns, animal welfare) but also by egoistic motives (Magnusson et al., 2003). Similar results were found in the case of another emerging organic market, i.e. Albania, Russia, and Pakistan. In the Russian organic market, organic food consumption is mainly motivated by personal well-being while social or environmental concerns are of low importance (Bruschi et al., 2015). Similar results were found in the study conducted by Imami et al. (2017) who showed that food safety override environmental concerns of organic food shoppers in emerging and less developed markets.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to obtain empirical understanding of organic food buyers in the context of emergent organic food market (i.e. Bosnia and Herzegovina) by using a segmentation approach.\nDesign/methodology/approach: A self-administrated online survey was carried out among organic food buyers (n=202) using the snowball sampling technique. Measurement items were mainly adapted from the prior studies.\nFindings: The authors analysed the heterogeneity of organic food buyers with latent class model. Four distinct latent classes (i.e. segments) of organic food buyers were identified. Those segments were named as enthusiastic social-seekers, enthusiastic moralists, hostile seldom shoppers, and hostile heavy shoppers.\nOriginality/value: Though the study was exploratory, the identified segments of organic food buyers can enhance our knowledge about differing characteristics of organic food buyers in the context of the country where the organic food industry is in the early stages of development. The findings of this study will give organic food producers and marketers a much better framework for making product, pricing, distribution and marketing communications decision. Moreover, the identification of organic food consumer profiles will provide an insight into how policymakers should tailor their public policy and strategies to expand the size of the organic food market.\n",
"role": "assistant"
}
] |
A hierarchical typology of scholarly information units: based on a deduction-verification study
|
[
"Information science and documentation",
"Concept analysis",
"Information concept",
"Information unit",
"Scholarly information",
"Theory of LIS"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: A classical definition of information science defines it as a discipline \"that investigates the properties and behavior of information, the forces governing the flow of information, and the means of processing information for optimum accessibility and usability\" (Borko, 1968, p. 3); The same definition further characterizes the discipline as \"concerned with that body of knowledge relating to the origination, collection, organization, storage, retrieval, interpretation, transmission, transformation and utilization of information.\" As far as the discipline's role with information is concerned, two features of this definition stand out and still hold true for today's library and information science (LIS). First, it states that the ultimate objective of the discipline is to ensure the greatest accessibility of information; second, it emphasizes that, to do so, the information profession performs a wide range of activities on information. Clearly, many of these activities are physical or quasi-physical operations that presuppose a certain degree of concrete tangibility of information and the possibility of dividing information into discrete operational units. This inevitably raises the question of what is meant by \"a unit of information\" in LIS.\n2. Review of previous research: Prior to the advent of digital technologies, library and information professionals relied very much on material documentary units (e.g. a book, a journal issue), as opposed to information units, to organize their activities. However, even then, pioneer documentalists such as Otlet and Kaiser began to investigate ways of identifying information units for indexing (Rayward, 1990; Dousa, 2010, 2013, 2014). As Dousa's (2013) historical study notes, in the heyday of documentation, special librarians and documentalists took pride in processing information rather than documents; the latter, as they believed, being the remit of general librarians. Their claim to work with information rather than documents and to be distinctive from general librarians naturally led them to consciously seek information units to operate with. For Otlet (in Rayward, 1990), demarcation of information units is an indispensable step toward developing his encyclopedia-like repertoire of human knowledge, for the development of this repertoire would necessarily involve extracting information units from their source documents, transcribing them onto retrievable sheets or cards and then rearranging them according to a certain order. For Kaiser (1911), it is a prerequisite for ensuring effective and efficient access to information in organizational contexts, as organizational tasks often demand direct access to relevant components within a document, in addition to accessing the document as a whole. Both pioneers advocated identifying information units according to their intellectual contents (facts or ideas). Kaiser (1911, section 73), for instance, identified information units according to the following contents: first, the concrete (e.g. a commodity) of which an information unit is about; second, the action of the concrete; and third, the locality with which the concrete and its actions are associated. Both pioneers also emphasized the exclusive correspondence between an index entry and the information unit it describes, contending that each individual unit should be recorded on a separate card or sheet, and each card or sheet should correspond to a definite unit. Otlet termed this correspondence \"the monographic principle\" (Otlet, in Rayward, 1990, p. 149).\n3. Research design: This study continues to address the issue of defining information units for LIS professional activities. It first derives a list of information units from a conceptual analysis of information and then subjects them to empirical tests via surveys of both information products and information users. The whole research process therefore consists of three procedures: a deductive enumeration of possible units, an empirical test of these units with a set of scholarly papers and, finally, an empirical test of the resulting units with a sample of information users.\n4. Deduced information units based on a conceptual analysis of information: As stated in the introduction section, this study is based on the assumption that to disaggregate information into units, it is mandatory to know what information is. Unfortunately, information has remained one of the most controversial LIS concepts until today. This study does not attempt to resolve all the controversies before examining information units; it takes one of the existing conceptualizations as the starting point instead.\n5. Results of the first empirical test: As reported in the methodology design section, in order to test the actual existence and detachability of the deduced information units in scholarly journal articles, this study invited 11 PhD candidates to mark all the chunks of their selected articles that they regarded as independently valuable for search and access. Marked chunks were then coded by two researchers using the information units in Table I. In total, the 11 participants returned 19 valid articles.\n6. Results of the second empirical test: As mentioned earlier, in addition to a close examination of actual articles, this study also conducted a questionnaire survey to ask information users whether they would like to have the possibility to retrieve and display the listed information units independently when searching full-text databases. The questionnaire contains only the segmental units thitherto obtained, assuming that a section containing any valuable segments will also be valuable itself. In total, 138 valid questionnaires were returned. The results are shown in Table III. As the purpose of this investigation is to test whether these units would be regarded as useful by at least some people in some contexts, user attributes were not taken into consideration.\n7. Conclusion, implications and limitations: 7.1 A hierarchical typology of scholarly information units\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to lay a theoretical foundation for identifying operational information units for library and information professional activities in the context of scholarly communication.\nDesign/methodology/approach: The study adopts a deduction-verification approach to formulate a typology of units for scholarly information. It first deduces possible units from an existing conceptualization of information, which defines information as the combined product of data and meaning, and then tests the usefulness of these units via two empirical investigations, one with a group of scholarly papers and the other with a sample of scholarly information users.\nFindings: The results show that, on defining an information unit as a piece of information that is complete in both data and meaning, to such an extent that it remains meaningful to its target audience when retrieved and displayed independently in a database, it is then possible to formulate a hierarchical typology of units for scholarly information. The typology proposed in this study consists of three levels, which in turn, consists of 1, 5 and 44 units, respectively.\nResearch limitations/implications: The result of this study has theoretical implications on both the philosophical and conceptual levels: on the philosophical level, it hinges on, and reinforces the objective view of information; on the conceptual level, it challenges the conceptualization of work by IFLA's Functional Requirements for Bibliographic Records and Library Reference Model but endorses that by Library of Congress's BIBFRAME 2.0 model.\nPractical implications: It calls for reconsideration of existing operational units in a variety of library and information activities.\nOriginality/value: The study strengthens the conceptual foundation of operational information units and brings to light the primacy of \"one work\" as an information unit and the possibility for it to be supplemented by smaller units.\n",
"role": "assistant"
}
] |
The impact of social media use for communication and social exchange relationship on employee performance
|
[
"Leader-member exchange",
"Social media",
"Employee performance",
"Horizontal communication",
"Team-member exchange",
"Vertical communication"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Social media such as Facebook, Twitter and WeChat, has increasingly gained organizations' attention (Papa et al., 2018; Scuotto et al., 2017; Song et al., 2019). It is estimated that social media can generate more than $1tn of the economic impact on business, and most of the economic value will be gained from more efficient collaboration and communication within and across organizations (Chui et al., 2012; Kane et al., 2014; Lu et al., 2015). Large numbers of organizations including Google, Microsoft and IBM have invested considerable resources on constructing social media tools to enhance collaboration and communication among employees with the expectation of improving their job performance (Lu et al., 2015; Song et al., 2019).\n2. Literature review: 2.1 Theory of virtuality\n3. Research model and hypotheses development: Based on the theory of virtuality,we argue that social media use for vertical communication (SMUVC) can positively influence LMX, and social media use for horizontal communication (SMUHC), can positively affect TMX. In addition, we further examine how LMX and TMX simultaneously enhance employee performance. Moreover, the moderating role of task complexity is also incorporated in the model. Figure 1 presents the research model and the corresponding hypotheses are elaborated below.\n4. Research methodology: 4.1 Sample and participants\n5. Data analysis and results: 5.1 Measurement model\n6. Discussion and implications: 6.1 Summary of results\n7. Conclusion: In recent years, researchers have realized the importance of social media to support communication among employees within organizations. However, there is a lack of research investigating how social media use for communication affects employee performance. Our findings suggest that social media use for vertical and horizontal communication exert great effects on LMX and TMX, respectively. In addition, LMX and TMX simultaneously enhance employee performance. Task complexity positively moderates the relationship between LMX and employee performance, while it negatively moderates the relationship between TMX and employee performance. Collectively, these interesting findings offer guidance for future research to further examine the role of social media in the workplace.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the underlying mechanisms through which social media use for vertical and horizontal communication enhance employee performance.\nDesign/methodology/approach: To test the research model, the authors conducted a questionnaire survey in China. The authors used a customer panel database provided by a marketing research firm in China to identify appropriate respondents. Finally, the authors received 243 valid responses.\nFindings: The authors find that social media use for vertical communication (SMUVC) is positively related to leader-member exchange (LMX) and social media use for horizontal communication (SMUHC) is positively related to team-member exchange (TMX). LMX and TMX are positively related to employee performance. LMX is positively associated with TMX. Besides, task complexity positively moderates the relationship between LMX and employee performance, while it negatively moderates the relationship between TMX and employee performance.\nOriginality/value: First, it adds to the literature by investigating the underlying mechanisms of how social media use for communication influences job performance. By identifying LMX and TMX as the underlying mechanisms, the authors make comprehensive considerations of how the vertical and horizontal relationships link the effect of social media use for communication on employee performance. Second, despite the growing evidence demonstrates that high-quality LMX and TMX can individually contribute to employee job performance, little research has considered both LMX and TMX relationships simultaneously and their effects on job performance. Finally, by establishing task complexity as a key moderator on the relationships between LMX and TMX and job performance, the study could explain the inconsistent findings in the literature that the effects of LMX and TMX are significant in some studies yet not significant in other studies.\n",
"role": "assistant"
}
] |
Behavioural economics and financial services marketing: a review
|
[
"Financial services",
"Behavioural economics",
"Consumer behaviour in financial services",
"Marketing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: It has been posited that insights provided by behavioural economics have profound implications for the study and practice of marketing (Mitchell, 2010). Equally, perspectives provided by behavioural economics have been advanced as effective in helping to understand consumer behaviour in financial services markets (Dixon, 2006; de Meza et al., 2008). In this paper, we seek to illustrate and explain how various insights offered by behavioural economics help enhance our understanding of marketing and consumer behaviour in financial services markets and to explore implications for firms, policymakers and other interested parties. Behavioural economics is an umbrella term for a range of approaches that seek to understand and explain observed consumer behaviour more accurately than predictions associated with traditional economic theory. Traditional economics is based on the assumption that people act in a completely rational and instrumental, or self-interested, manner. People are assumed to take full account of all information available to them and to act according to this information and their preferences. In doing so, their goal is assumed to be maximisation of their utility. Standard economic models incorporate such assumptions, as, by implication, do any marketing strategies or regulations and policies based on such models. For instance, recent approaches to consumer policy in the UK are effectively predicated on the notion of the rational consumer. The working hypothesis has been that provided consumers are supplied with as much information as possible and efforts are made to improve their financial capability, then consumers should make more informed and suitable choices in financial services markets.However, it is increasingly apparent that rational, \"information-processing\" models do not accurately explain actual or observed consumer behaviour. An increasing body of evidence points towards persistent and consistent violations of outcomes expected according to a rational, or information processing, perspective (Thaler and Sunstein, 2008). These violations occur for various reasons, including the mental effort required to make informed decisions, faulty reasoning on the part of consumers and the complexity of the decision-making process. From an economic perspective, such observed behaviour would be regarded as irrational and would be characterised as anomalous. However, as such behaviour is observed consistently and the evidence is relatively robust, departures from what economists consider to be rational need to be recognised and explained, rather than ignored or discounted as mere aberrations. Such alternative models of consumer behaviour also need to be accounted for by financial services policymakers and by firms seeking to influence consumer choice in financial services markets.It is perhaps not surprising, therefore, that there has been growing interest in the subject of behavioural economics, a relatively recent addition to the study of economics which tries to explain such anomalies with reference to consumer psychological insights. Increasingly, it is recognised that we cannot study human behaviour as a \"black-box\" (i.e. not concerned with the process of behaviour merely the outcomes as predicted by economic models), we need to open the box and look into the cognitive processes that drive human behaviour. The objective of behavioural economics is to bring psychological insights into economics in a systematic manner, in order to improve economic theory, so that it can better explain human behaviour. In doing so, the impact of the typical mental short-cuts, or heuristics, that consumers employ in their decision making are taken into account, as are other systematic biases, anomalies and framing effects that impact on consumer choice and behaviour.This paper considers a number of important insights from behavioural economics which may be of interest and use to marketers and policymakers in the financial services sector. Our objective is to review and apply to a financial services context a number of salient insights offered by behavioural perspectives. It would not be practical to attempt to be all-encompassing in our coverage. Rather, we choose to focus on elements of the behavioural economics approach which we think provide the most important and useful insights. Implications for those responsible for formulating consumer policy in financial services receive particular attention as it is in this area that the potential impact of behavioural perspectives is particularly profound. We investigate loss aversion, status quo bias, framing effects, hyperbolic discounting and mental accounting, all of which deal with biases in the judgment of alternative outcomes that arise due to the way these outcomes are presented, appraised and compared. We also detail the availability effect which is an example of how biased information collection and processing results in erroneous beliefs. Finally, we look at overconfidence and trust, which are factors besides utility maximisation that may influence the selection between alternatives, giving rise to actions which would be considered erroneous from the standpoint of traditional economic theory. We have chosen these particular facets of behavioural economics as we feel that they provide particularly important and salient insights in the context of financial services. In this paper we focus on implications for interactions with financial consumers, rather than dealing with wholesale markets and macro-economic matters. In the main section of the report, the various phenomena introduced above are introduced and explained briefly, before examples from a financial services context are provided.\nBehavioural economics and consumer behaviour in financial services: Loss aversion, prospect theory, the disposition effect and the endowment effect\nConclusions: The insights afforded by behavioural economics are of great interest to policymakers and practitioners in financial services. An understanding of the implications of behavioural economics for elements of consumer behaviour may help in fashioning a choice architecture that is more likely to bring the desired consumer responses. From a commercial perspective, the objective is to use the lessons provided by behavioural perspectives to enable companies to present choices in a manner that maximises take-up of products and services, encourages or engineers ongoing behavioural loyalty and thereby helps to cement relationships between firms and customers. From the policymakers' perspective, insights from behavioural economics may be used to encourage consumers to engage with financial services to a greater degree and to make increased levels of provision in areas such as pensions and other saving and, while doing so, to make ongoing investment decisions which are in the consumer's long-term interests. This paper has provided an explanation of a number of aspects of behavioural economics and has supplied numerous examples of how these heuristics, biases and mental shortcuts may impact on financial services products and markets. In particular, the importance of loss aversion, the status quo bias and defaults and framing effects have been highlighted.\n",
"role": "user"
},
{
"content": "Purpose: - It has been argued that the insights provided by behavioural economics have profound implications for the study and practice of marketing. The purpose of this paper is to provide a detailed analysis of how such insights help enhance understanding of aspects of marketing and consumer behaviour in financial services markets.\nDesign/methodology/approach: - This paper looks at various facets of behavioural economics which it is felt provide particularly important and salient insights in the context of financial services. In particular, it studies loss aversion and prospect theory, status quo bias and defaults, framing and anchoring effects, hyperbolic discounting, availability effect and salience and over-confidence. In doing so, it provides a number of examples from the financial services context which provide insightful and informative insights for both commercial practitioners and policymakers. In each case, relevant phenomena are introduced and explained before providing a number of applications and examples from the financial services domain.\nFindings: - The authors find that the insights afforded by behavioural economics are useful in helping to explain various aspects of consumer behaviour in financial services markets. It is shown that an understanding of the implications of behavioural economics may help in fashioning a choice architecture that is more likely to bring the desired consumer response, from either a commercial or policymaking perspective.\nPractical implications: - This analysis provides important insights for those responsible for the marketing of financial services, policymakers in the financial services domain, third sector agencies seeking to foster greater engagement with financial services and other interested parties.\nOriginality/value: - The paper adds value by drawing together various aspects of behavioural economics, providing an analysis of their relevance to financial services marketing and offering numerous examples and applications.\n",
"role": "assistant"
}
] |
Strategic management accounting of social networking site service company in China
|
[
"China",
"Business formation",
"Management accounting",
"Social networking sites",
"Popular culture"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: We aim at providing real-time as well as non real-time users with a variety of products and services, and thereby steadily increasing user loyalty (CEO of Tencent) (Ma, 2010).Tencent QQ (QQ) has become interpersonal relationship managing software, administrating the social contacts between a countless number of people (CTO of Tencent) (Ma, 2010).This paper studies a major social networking site (SNS) services company in China. It is a case study of a renowned China information technology (IT) company, Tencent, which provides SNS services, and it analyzes the roles of strategic management accounting (SMA) in relation to organizational life cycles.Companies in both the manufacturing and service sectors have faced dramatic changes in their business environment (Ma and Tayles, 2009; Langfield-Smith, 2008). Increase in government regulation, combined with extensive competition from overseas companies in domestic markets has led to a highly competitive global market. It influences management to combine traditional management accounting techniques with business strategies in operation and consequently leads to emergence of SMA. SMA introduces a long-term and external perspective, and it emphases the opportunity to rationalize the accounting information used for building the business strategy. The origin of it dates back to Simmonds (1981, 1982). He developed the concept of SMA, stating it as a collection of management accounting information about the business and its competitors, which are used in developing and monitoring the organisation's strategy. Regarding corporate growth, SMA extends to company finance spectrum for coordinating management strategy, accounting and finance (Langfield-Smith, 2008).SNS is a web-based service that enables users to create profiles, articulate and communicate with others. It is a fancy and popular cultural issue. Boyd and Ellison (2007) define SNS as a web-based service that enables people to:* construct a public or semipublic profile within a bounded system;* communicate with other users; and* view the pages and details provided by other users within the system.The SNS has evolved as a combination of personalized media experience, within social context of participation. As the increase in popularity of SNS is on a constant rise, new uses for the technology are constantly being notified. At the forefront of emerging trends in SNS is the concept of \"real time\" and \"location based\". It allows users to contribute content, which is then broadcasted as it is uploaded. The concept is similar to live television broadcasts. Nowadays, SNS is a popular culture throughout the world. Popular culture incorporates series of ideas, perspectives, attitudes and other phenomena that are deemed an informal consensus within the mainstream of culture (Oracle ThinkQuest, 2000). Heavily influenced by the mass media, this collection of ideas reflects the everyday lives of the society. Popular culture changes constantly and occurs uniquely in place and time. SNS represents a complex of mutually interdependent perspectives and values that influence the society.SNS allows users to create profiles, personal demographic data, for themselves; examples are Facebook, Twitter and Tencent QQ. SNS is openly available to all web users to communicate and is designed to attract online advertisements. In China, Tencent QQ has more than 400 million users, just ranking second after Facebook of 500 million users in mid-2010 (Tencent, 2010b).In China SNS market, Tencent's QQ, Baidu, Kaixin 001 and RenRen are the most popular SNS titles, according to an online brand recommender, Holaba (Kelikuru.com, 2011). For popularity, shopping frequency and recommendation, Tencent's QQ surpasses all other three competitors in China. QQ is now the most popular free SNS and instant messaging (IM) computer program in China. Since its entrance into the Chinese households, QQ has quickly emerged as a modern cultural phenomenon, now being portrayed in popular culture. Aside from the online chat program, QQ has also developed many sub-features including games, virtual pets, and ring tone downloads (Appendix).The development of this paper is structured as follows: the next section reviews the literature on SMA, with Section 3 integrating the literature on SMA and company life cycle. It is followed by development of three research propositions. Section 4 illustrates the research model. Section 5 provides the company's background and a detailed description of the introduction of SMA in the start-up and growth stages of Tencent. Section 6 discusses the findings of the research. Finally, Section 7 concludes the paper and discusses the limitations of the research and explores the areas for future research.\n2 SMA and social network: 2.1 Strategic management accounting\n3 Organizational life cycle and SMA: There has been abundant literature to discuss company development in relation to different life cycle stages (Greiner, 1998; Kazanjian, 1998; Miller and Friesen, 1983, 1984). These stages are commonly categorized as start-up, growth, maturity and decline. Since the incorporation of Tencent in autumn 1998, it has successfully passed through the start-up stage and is currently in the growth stage.3.1 Start-up stage\n4 Research methods: The research methods are based on a case study approach for investigating into the three propositions. Yin (1994) stated that the case study is useful to investigate on a phenomenon that is not completely studied. SNS is one emerging popular culture service while Tencent is the largest and most significant SNS service provider in China. Thus, a single case study using Tencent is appropriate to analyze the issues highlighted in the previous section.In the current study, data utilized are mainly the archival data from company reports and newspaper clippings collected during July-September in 2010 and updated in early March 2011. Hong Kong Stock Exchange collects archives of listed company annual reports, and web services like Baidu.com collect archives of newspaper clippings of companies. The former provide official data and information about the listed companies in Hong Kong. The newspaper clippings collected through Baidu search engine provides further data and information interpretation and analysis from various viewpoints.Whitlow (2004) stated advantages of archival research include data already collected. They need not require further verification, for instance, the published company annual report has already been audited. Moreover, data may allow looking at variables that would be difficult or impossible to study directly, such as the company life cycle development in the current study.\n5 Tencent: Tencent is a publicly owned holding company which provides internet and mobile phone value-added services and operates online advertising services in China. It was found by a Computer Engineer, H.T. Ma, in November 1998 in Shenzhen, China. Tencent's diverse services include web portals, e-commerce, games, and its well-known SNS service, Tencent QQ (QQ). It runs one of the largest web portals in China, QQ.com. From 2004 to 2009, the profit has grown by 11.8 times (Tencent, 2009a, b). Just for the second quarter of 2010, the quarter profits rose 60.7 percent year-to-year to USD281.8 million, driven by extremely strong revenue generated virtual goods sales in its various online games (Tencent, 2010b).QQ was initially developed by Tencent in February 1999. It was first released as a Network Paging real-time communications service. Then other features were successively added such as chatrooms, games, personal avatars, internet storage, and internet-dating services (Appendix). Most of these features were free. Its variety of features and easy-to-use interface made it very popular, especially among teenage users. QQ is the most popular free IM computer program in China, and the number of simultaneous online QQ users exceeded 80 million (Tencent, 2009a, b). Since its entrance into Chinese households QQ has quickly emerged as a modern cultural phenomenon, now being portrayed in popular culture. Aside from the chat program, QQ has also developed many sub-features including games, virtual pets, and ring tone downloads (Appendix). Just in several years, QQ has become the most popular real-time communication service in China (CINIC, 2009). In less than ten years of incorporation, it has rapidly grown to have a market value of USD14.7 billion (Tencent, 2009a, b).The CEO, H.T. Ma, said, \"For me, now QQ is not the work, it is the interest\" (Ma, 2010). Like many other IT developers, he developed the IT service mainly basing on his interest. He added, \"Makes the company facing the shareholder, the management should be responsible to investors.\" He well accepts the idea of corporate responsibility to shareholders.Tencent revenues from internet value-added services, mobile and telecommunications value-added services, and online advertising business increased by 62.2 percent to USD686.6 million for the second quarter of 2010 from USD423.3 million comparing to the second quarter of 2009 (Tencent, 2010b). This primarily reflected the continuous pursuit of QQ as a popular culture item, which resulted in the increased monetization of company major advanced casual games. At the same time, the cost of revenues increased by 58.7 percent to USD218.07 million for the second quarter of 2010 from USD137.4 million for the second quarter of 2009. This mainly reflected the increase in sharing costs, staff costs, equipment depreciation and telecommunications operators' revenue share. All these are favourable financial data that reflect significant increase in profit.5.1 Start-up stage\n6 Discussion: 6.1 Proposition in start-up stage\n7 Conclusions: Tencent continues to enhance the user value of SNS platforms by improving the basic social-networking functionalities as well as introducing different SNS applications that cater to the varying needs of user base. Tencent also increases effort to extend QQ to wireless platforms to further enhance its coverage and usage.The evidence presented shows that Tencent has spent time and efforts in applying cost-leadership and product-differentiation strategies in the start-up stage, and the focus was on market and not on cost of operation. These strategies were successfully fulfilled using the appropriate internal accounting systems for administering company costs. Later, in the growth stage, financing channels as stock listing and trade financing were introduced. It raised shareholders' funds through stock listing in Hong Kong. Tencent coordinates well-corporate trading funds through the two online services, Tenpay.com and Paipai.com. Through the SNS services provided, the company started to become more external resources and opportunity focused in decision making. It emphasizes on strategic opportunities learnt from the experiences of other popular culture developers, such as Facebook.com, and implements this IT-based strategic business in providing SNS services.The single case study is a valuable research method when used for the purpose of analyzing how people frame and solve problems (Barzelay, 1993). This case study of Tencent provides specific investigation of how H.T. Ma developed and uplifted SNS service as QQ in China. It illustrates and analyzes reasons why different accounting and finance techniques are used by the company in different stages of the organizational life cycle. Nevertheless, the paper is based on a single case study, thus the findings may not be generalized to other popular culture service companies. Still this case study is able to show the interaction in different organizational life cycle stages. For example, Tencent emphasized much on its charged value-added services to compensate for the operation cost of free membership provided in the start-up stage. It could even be deduced that the company utilized this low-cost strategy to expand the customer base. Even in the later growth stage when the company raised funds in financial markets and used them in funding operation, there was still an explicit link back to free membership strategy to maintain its market share. The new financing means did not undervalue the importance of free membership for maintaining market share.From a practitioner standpoint, the findings of this study show that one way to become a successful SNS service company for promoting popular culture is to develop good SMA techniques for matching development needs of different organization life cycles and then to design appropriate business strategies matched with management accounting.Future research is suggested to base on this initial study to conduct a more in-depth study in which SMA is implemented in high-growth companies engaging in popular culture. This can be done through a survey of low growth verses high growth companies to examine the different ways in which the companies implement SMA techniques and systems in different stages of organizational life cycles.\n",
"role": "user"
},
{
"content": "Purpose: - Social networking site (SNS) service is a popular cultural issue that affects communication methods and mannerisms. The purpose of this paper is to analyze the case of the growth and development of a China SNS service company using the strategic management accounting (SMA) spectrum.\nDesign/methodology/approach: - The paper primarily employs archival resources such as company reports and newspaper clippings online to provide a business and social developmental context to the study. Three propositions are developed and discussed.\nFindings: - The study illustrates SMA application in business. It shapes the perception of contemporary accounting, and then delivers an awareness of the relationship among business strategy, accounting, and social life.\nResearch limitations/implications: - This paper is based on a single case study, thus the findings may not be generalized to other popular culture service companies. The case has shown some reasons why different SMA techniques are used by the company in different stages of the organizational life cycle.\nPractical implications: - First, in the start-up stage of an SNS service company, it engages in product development to obtain funding resources and market so as to enhance market share. Second, in the growth stage of SNS service, the company engages in product financing through the capital market and trade financing.\nOriginality/value: - The SNS service has emerged for accounting for popular culture. Such connection is especially found in SMA in different stages of the organizational life cycles as addressed in the study.\n",
"role": "assistant"
}
] |
A panel for lemons? Positivity bias, reputation systems and data quality on MTurk
|
[
"Methodology",
"Data quality",
"Online research",
"Positivity bias",
"Reputation systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Since its introduction in 2004, Amazon's Mechanical Turk (MTurk) has become an essential tool for marketing researchers, enabling the rapid collection of data from thousands of participants worldwide. With 15,000 papers referencing MTurk published between 2006 and 2014 (Chandler and Shapiro, 2016), the importance of this tool, along with other markets for crowdsourced work such as CrowdFlower and Prolific Academic (Peer et al., 2017; Vakharia and Lease, 2015), is difficult to overstate. As the popularity of these crowdsourcing panel systems has grown, a considerable body of work has emerged characterizing the nature of respondents and codifying best practices in the panels' use. A cornerstone of these recommendations is an emphasis on methods of ensuring high quality in the data that workers produce. Though some recent research has investigated the prevalence of various forms of worker deception on MTurk (Chandler and Paolacci, 2017; Sharpe Wessling et al., 2017; Suri et al., 2011), for many academic research paradigms, the quality of work is more subjective.\nEnforcing cooperation in online transactions: In online markets, where participants interact without face-to-face contact, systems that enable cooperation by creating trust and ensuring good conduct are necessary. In these transactions, the fundamental problem is one of incomplete and asymmetric information: typically, one party (the seller) possesses private information about their type or effort level and produces high- or low-quality work, while the other party (the buyer) has a preference for high quality and wishes only to do business with those who provide high-quality output (Akerlof, 1970). When the quality of the seller's product is hard to assess until some point after purchase, such as with an experience or credence good (Gao et al., 2015; Kokkodis and Ipeirotis, 2015; Nelson, 1970), this information asymmetry creates the possibility for two negative outcomes for the marketplace: adverse selection and moral hazard. With adverse selection, low-quality sellers attempt to conceal their type from buyers, allowing them to charge higher prices, while moral hazard occurs when there is no incentive for high-quality workers to maintain quality, and they engage in opportunistic behavior by shirking (Pavlou et al., 2007). With both of these possibilities, buyers receive products of lower than expected quality, which undermines trust and may lead buyers to exit entirely.\nReputation and litigation on Mechanical Turk: Because of the effectiveness of both reputation and litigation in ensuring cooperation, they have become indispensable tools for online transactions, and both feature prominently on MTurk. In the research market on MTurk, researchers, as MTurk requesters, act as \"buyers\" of labor (an experience good, Kokkodis and Ipeirotis, 2015) from \"sellers,\" i.e. MTurk workers. To facilitate these transactions, MTurk provides a reputation system and a system akin to litigation, in the ability to accept or reject work. The reputation system automatically tracks several measures that are referred to as \"system qualifications.\" The two primary qualifications are productivity, measured as the number of tasks a worker has completed, and the approval rate, measured as the percentage of the overall number of tasks completed that have been approved by requesters (Sheehan and Pittman, 2016). MTurk's reputation system also measures workers' \"Masters\" status, who are identified by MTurk as \"elite groups of Workers who have demonstrated accuracy on specific types of [tasks]\" (Amazon, 2018). However, the criteria under which workers receive this status is unclear (Yin et al., 2016), beyond having completed 1,000 HITs and maintaining a 99 per cent approval rating (Clickhappier, 2016). Thus, being granted the Masters qualification confers a status on workers that distinguishes them and their quality in a manner that is separate from their reputation (Azoulay et al., 2013; Malter, 2014; Simcoe and Waguespack, 2011).\nStudy 1A: The purpose of the Study 1A was to provide evidence for the potential skew in worker reputations on MTurk, by parameterizing its distribution. Though prior work has provided estimates of average approval rates on MTurk of approximately 99 per cent (Schulze et al., 2013; Yin et al., 2016), these results are limited for two reasons: first, they rely on self-reports and thus may be subject to potential misrepresentation (Chandler and Paolacci, 2017), and second, they do not provide insight into the values that are actually observed by requesters on the MTurk platform. These values are the most relevant for researchers, as they are observable prior to data collection and can be used to limit participation to workers who meet the requester's criteria.\nStudy 1B: To extend these findings by showing the relationship between what is observable to researchers and true worker characteristics, a survey was conducted. Four separate HITs containing three questions were posted simultaneously on a weekday at 1:00PM UTC, each recruiting 40 USA participants who had completed at least 100 HITs. Similar to Study 1A, each HIT had a different approval rate requirement, of exactly 97 per cent, 98 per cent, 99 per cent or 100 per cent. In the task, participants were asked to enter the total number of HITs they had submitted, and the number of HITs they had worked on that had been approved and rejected as reported in their worker dashboard.\nStudy 2: The purpose of Study 2 was to test conditions consistent with moral hazard. The assumption is that if all workers are of high quality, then in the presence of a positivity bias, workers may engage in shirking unless they are incentivized not to do so. A potential way to demonstrate this is to use repeated interactions in a panel design, because the awareness of future opportunities reduces incentives to defect (Kim, 1996). Thus, if workers believe their performance may affect their ability to work on a future task, they should be less likely to engage in behaviors that produce low-quality responses. However, a key consideration of this approach is establishing awareness of this performance contingency, but not of the precise criteria (Sharpe Wessling et al., 2017). Therefore, two different recruitment procedures were used for the panel conditions. In the first, at the beginning of the initial interaction, participants were informed about a subsequent study and that invitations were contingent upon performance. In the second, this was not mentioned. Responses in these two conditions were compared against each other, as well as against a one-shot condition where the entire interaction would be completed in a single HIT. If shirking is present, this should manifest in higher quality responses when participants are aware that the possibility of completing future jobs depends upon their performance.\nStudy 3: The purpose of the third study was to examine the second scenario, where worker differentiation is present, but the reputation system is uninformative about these differences. The focus was on the two primary measures of the reputation system - HIT approval rates and productivity, along with a status marker in the Masters qualification. In addition, conditions were included to compare the effects of payment on performance.\nStudy 4: The goal of Study 4 was to examine the final scenario, which predicted that the skewed distribution was the result of a positivity bias, but that information was present in the extreme levels of the reputation system. Based on the results in Study 1, the recommended reputation level threshold of 95 per cent effectively includes more than 98 per cent of the workers on MTurk, and this is only 9.4 per cent more of the total worker population compared to the group of workers with approval rates of 99 per cent or higher. Thus, it is reasonable to think heterogeneity in quality exists between the discrete levels of the reputation system above 95 per cent approval rates that was unobserved in Study 3.\nGeneral discussion: The purpose of this research was to investigate how the merging of reputation and litigation mechanisms into a single tool, along with the presence of a positivity bias, can create conditions for a reputation system to fail to provide quality sorting. In five studies exploring this phenomenon on Amazon's MTurk, evidence was presented that the distribution of reputation is highly skewed in a manner consistent with a positivity bias, and that there is heterogeneity in quality among workers that is captured by status markers and at extreme values of reputation. While the findings did support the prediction that there can be adverse selection unless these are accounted for, no evidence was found for moral hazard, suggesting that workers on the platform were sufficiently motivated to limit shirking.\nConclusion: Careful consideration is necessary when creating systems to ensure cooperation in marketplaces. The findings of this work suggest the design of MTurk's system introduces biases into evaluations, which affects the infromation it contains about worker quality. In practice, the presence of these biases means that most requesters should raise their approval rate thresholds to 99 per cent, well above those recommended by most literature on MTurk and commonly employed by researchers. Further, it is important for researchers to be mindful of their obligations to help maintain the panel, by accounting for these biases.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate how the effectiveness of systems for ensuring cooperation in online transactions is impacted by a positivity bias in the evaluation of the work that is produced. The presence of this bias can reduce the informativeness of the reputation system and negatively impact its ability to ensure quality.\nDesign/methodology/approach: This research combines survey and experimental methods, collecting data from 1,875 Mechanical Turk (MTurk) workers in five studies designed to investigate the informativeness of the MTurk reputation system.\nFindings: The findings demonstrate the presence of a positivity bias in evaluations of workers on MTurk, which leaves them undifferentiated, except at the extremity of the reputation system and by status markers.\nResearch limitations/implications: Because MTurk workers self-select tasks, the findings are limited in that they may only be generalizable to those who are interested in research-related work. Further, the tasks used in this research are largely subjective in nature, which may decrease their sensitivity to differences in quality.\nPractical implications: For researchers, the results suggest that requiring 99 per cent approval rates (rather than the previously advised 95 per cent) should be used to identify high-quality workers on MTurk.\nOriginality/value: The research provides insights into the design and use of reputation systems and demonstrates how design decisions can exacerbate the effect of naturally occurring biases in evaluations to reduce the utility of these systems.\n",
"role": "assistant"
}
] |
Perceptions of Chinese traditional food and European food among Chinese consumers
|
[
"China",
"Word association",
"Consumer",
"European food",
"Perceptions",
"Traditional food"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: China has the world's largest population, is experiencing sustained growth in personal income and has a dietary consumption pattern inclined towards Westernisation (Pingali, 2007; World Bank, 2014). This has led to a dramatic growth in demand for European food products (e.g. wine and beer) in this largest East Asian country recently (Alinna, 2013; Lu, 2014; Balestrini and Gamble, 2006). An increased importance has therefore arisen for the research areas in relation to consumers' attitudes, perceptions and behaviours towards European food products in East Asia, especially in China. A few studies do identify Chinese consumers' impressions of Western local food (too sweet, low in fruit and vegetables, high in calories, with too many uncooked or cold dishes and too much fried food), as well as the factors influencing their preferences for Western foods (taste, price, hygiene, safety, appetising assurance or familiar flavour) (Chang et al., 2010; Curtis et al., 2007; Li et al., 2011). However, there is still a lack of understanding with regard to Chinese consumers' perceptions towards European food; how do Chinese consumers define the concept \"European food\"?\nMaterial and methods: Web-based free word association test\nResults and discussion: Grouping the elicited words into classes\nConclusions and limitations: By using an online qualitative research method (word association test), this study is one of the first to address how Chinese consumers define both European food and their own traditional food in China, the largest East Asian country. In summary, Chinese consumers have clear perceptions of the two food concepts, with 13 dimensions emerging for the concept of Chinese traditional food and 12 dimensions for the concept of European food. Although ten dimensions are the same, obvious differences were identified by comparing the classes for the two food concepts. For example, Chinese consumers' Health perceptions of Chinese traditional food were more positive than Health perceptions of European food. By contrast, their Safety perceptions of European food were more positive than for Chinese traditional food.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore Chinese consumers' perceptions in relation to both Chinese traditional and European food.\nDesign/methodology/approach: A web-based free word association test was administered to 302 consumers in China. They were asked to give the first three words that came into their minds when they were presented with each of two stimulus words, \"traditional food\" and \"European food\". Three researchers grouped the elicited words into classes and then into dimensions. kh2 association tests were used to assess and identify statistically significant differences in the frequencies of classes and dimensions for the two food concepts between socio-demographic groups.\nFindings: The findings show that Chinese consumers define Chinese traditional food and European food through ten similar dimensions: Sensory appeal, Health, Origin, Marketing, Safety, Variety, Heritage, Symbolic meaning, Simplicity and Mood. Additionally, they associate Chinese traditional food with the dimensions Elaboration, Celebration and Habit, as well as link European food to the dimensions Convenience and New. Although ten dimensions are the same, obvious differences can be identified by comparing the classes for the two food concepts. Further, there are significant differences in the class associations for European food between age groups and in the dimension associations for Chinese traditional food between gender groups.\nOriginality/value: By using an online qualitative research method, this study is one of the first to address how Chinese consumers define both European food and their own traditional food in China, the largest East Asian country. The findings are particularly useful for the future development of traditional food products and for the future export of European food products onto China and even other countries in East Asia.\n",
"role": "assistant"
}
] |
How do ideas and discourses construct youth policies? The case of Hong Kong
|
[
"Hong Kong",
"Ideas",
"Discourses",
"Documentary study",
"Youth policies"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: While youth policies are diversely practised and understood under various welfare regimes, political economies and cultural contexts, they are typically subsumed into different fields of policies with a blurred boundary (France, 2016). These complexities and ambiguities therefore lead to an added challenge for studying youth policies as a set of state interventions and a field of study. Inspired by recent scholarship of ideational and argumentative (re)turn in policy studies and political science, the role of ideas in constructing and understanding policy development has been taken into account in explaining policy continuities and changes. For decades, the use of ideas and discourses as analytical tools are criticised in the way that they reinforce the distinction between material and discursive perspectives in a de-contextualised and de-materialised manner. Through the new lens of the broad ideational school of institutionalism, ideas are unavoidably shaping and shaped by policy context and content. Therefore, it is of vital importance to study whether and how ideas and discourses matter when grounding and justifying policy arrangements in relation to the empowerment of policy actors. To avoid using voluntarist and reductionist approaches to examining ideas and discourses, the ideas scholarship reconnects ideas and institutions in order to go beyond an \"idealistic\" understanding of policy development. Drawing on this theoretical perspective, policy ideas are further manifested by a set of narratives, frames and beliefs that are taken as a common sense, and as the vehicles for expressing policy content and political justification (Schmidt, 2010). While the ideas scholarship has been applied into studying different fields of policies, both on local and international levels, the study of youth policies is however still under-researched. Youth policy could be understood as a set of state interventions, primarily focussing on youth-related issues, which is implemented under particular political economy and ideologies within the social policy domain (King, 2016). As a result of institutional arrangement negotiated and operated at the micro, national, and international levels, youth policy, as a field of practice (France, 2016), in general defines the meanings of 'being youth', constructs some \"problematizing\" assumptions of the younger generation, and shapes the everyday life and transitions of young people (Williamson, 2002). Youth development has gained increasing prominence as a movement and as a unifying theme for shaping youth policy and practice (Pittman et al., 2000), which recognises the agency and strengths of youth and emphasises active organisational, institutional and policy support for their growing capacity and competencies in order to promote the development of youth and society as a whole (Hamilton and Hamilton, 2004). Yet, how youth development is to be defined, constructed and promoted is also subject to a wide range of discussions and actions in the ecology of youth policy, which is shaped by political and ideological forces and policy delivery (see France, 2016).\nInsights of the ideas scholarship into the study of youth policies: Whilst ideas as a subject of research is nothing new in policy and youth studies, there is a growing body of literature revisiting the ideational, argumentative and discursive turn in political analyses and policy studies, such as the theories of institutionalism in particular (Beland, 2009; Howarth, 2010; Schmidt, 2010; Hay, 2011; Fairclough, 2013; Jessop, 2013; Blyth, 2016). It is argued that the three \"old\" strands of neo-institutionalism have different logics of explanations for institutional changes. The strands of rational institutionalism, sociological institutionalism, and historical institutionalism are driven by the logic of calculation, appropriateness, and path dependence, respectively. Despite their strengths in locating institutional continuity, they may lead to a static analysis and an equilibrium-centred theory, and thus downplay the role played by ideas and agenda-setting processes (Beland; 2009; Schmidt, 2015). To capture the nature and features of ideas, and their constructive power in politics and policy making, some argue to have an ideational approach to institutionalism, as the fourth strand of institutionalism, in order to achieve a more dynamic explanation for institutional changes with regard to agential perspectives (Campbell, 2002; Fischer, 2003; Beland, 2009; Hay, 2011; Rayner, 2015; Schmidt, 2015). Generally, policy ideas can be classified into two categories: cognitive and normative (Campbell, 2002; Schmidt, 2010). Instrumental interest is the driving force behind cognitive ideas while moral value is the one behind normative ideas. In other words, the ideational school of institutionalism shares the \"logic of calculation\" and the \"logic of appropriateness\" to understand the content of policy ideas. Meanwhile, the justification of ideas is defined as the policy discourses that are driven by the \"logic of communication\", in which narratives and frames are disseminated and negotiated in a dynamic manner. Both of the cognitive and normative ideas upheld by the government embrace the problem definition of and solution selection for continual governance (Barbehon et al., 2015). It is also argued that discursive institutionalism explains the relationship between discursive activities and institutional context, which results in policy continuities (Hope and Raudla, 2012).\nChanging youth narratives and policies under the HKSAR's hybrid regime: The emergence of youth narratives and policies does not appear in a vacuum but is rooted in the domestic institutional arrangements. First, the HKSAR's political structure is perpetually depicted as a \"hybrid\" regime (Fong, 2013; Wong, 2015; Cheng, 2016) in which the polity is characterised by liberal and competitive authoritarianism. On the one hand, civic liberties are generally enjoyed by Hong Kong citizens, such as the freedom of speech and demonstrations. On the other, only half of the seats in the Legislative Council are directly elected by Hong Kong citizens through election as \"geographical constituency\"; therefore the opposition parties cannot win the majority power over the pro-China and pro-establishment parties (Cheng, 2016; Ng, 2016). Although electoral democracy is partially practised, the real power of governance is centralised in the hands of bureaucrats, economic elites and pro-government parties. They unite together and become a governing alliance in both legislative and executive institutions. Furthermore, the power of main officials, especially the CE, is secured by the basic law to formulate and implement policies in the executive-led regime. Under the HKSAR's non-democratic state, decision making and agenda setting in the policy processes are exclusive that the civil society is inadequately engaged (Lee et al., 2013). There is no independent bureau or department assuming the overall responsibility for youth policies in Hong Kong. Moreover, the Commission on Youth as an official consulting body to the HKSAR Government does not possess any statuary power to formulate and implement youth policies. Thus, there is no incentive for the HKSAR Government to institutionalise the consultation and documentation of youth policies which are used to explicitly identify the policy components that contribute to the welfare and development of youth. Two implications prevail: first, although the state interventions on young people are not justified as a coherent set of youth policies or are not placed under the administration of a youth bureau or department, youth policies de facto exist in a fragmented manner; second, as a powerful policy actor, the CE's ideas and discourses on the young are influential and indicative of youth policies in the HKSAR's hybrid regime. In other words, youth policies in Hong Kong can be scrutinised by examining the CE's notions of youth.\nMethods: Documentary study is used to map out the CE's fragmented ideas and discourses for several reasons. First, although there are no specific policy documents for youth policies in Hong Kong, the CE's ideas and discourses are officially recorded in various types of online documents available to the public. Second, there are technical difficulties in reaching the CE in order to study his views on youth policies and the reliability of using interviews is questionable due to political sensitivity. Third, documentary analysis can go beyond a sketchy and positivist understanding of policies by exploring the explicit information, rhetoric of policies, and underlying ideologies (Abbott et al., 2004). Hence, documentary study as a research method to examine policy complexities is not restricted to formality of texts. Finally, a documentary analysis of youth policies can inform other kinds of methods for the next stage of research, such as in-depth interviews with young people and other policy stakeholders.\nNormative narratives of behavioural patriotism in youth policies: Recent studies on youth development and youth policies reveal that negative portrayals of youth are no longer dominant in neoliberal governance agendas. In contrast, rather than simply outlining the negativity and problems of youth, the ruling bodies and policy actors have been visioning more positive development of young people by means of promoting new moralities and value judgements aligned with changing neoliberal contexts (Sukarieh and Tannock, 2011). Therefore, not only are problems and deficits identified in the governing ideas of youth development, but also the positive attributes and ideal characters of youth.\nCognitive notions of opportunistic economism in youth policies: According to the ideational institutionalism approach, interest per se is constructed and reframed in different social contexts. This contextualised cognitive approach to interest and ideas also goes beyond a mechanical perspective of rational institutionalism which assumes rationality and calculation as the ultimate driving forces of institutional change and continuity.\nConclusion: It is argued that the perspective of ideational and discursive institutionalism can inform policy studies by addressing the role which ideas and discourses play in the promotion of policy changes or the reproduction of continuities. The introduction of power as a dimension to analyse policy ideas further reconnects policy actors with policy institutions, in which the influences of social contexts upon ideas and actors are taken into account. Hence, ideas can both constrain and empower the political capabilities of actors. This dynamic relationship provides new insights into the studies of policy processes, policy content and policy contexts. In methodological terms, this analytical approach can be compatible with the qualitative approach to studying documents and interviews through developing different themes derived from the two types of ideas - normative and cognitive - categorised at three different levels, i.e. policies, programmes, and philosophies (Schmidt, 2010).\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to critically examine the youth narratives of Mr Leung Chun-ying, the Chief Executive (CE) of the Hong Kong Special Administrative Region (HKSAR) Government between 2012 and 2017, which steered the directions of youth policies in Hong Kong.\nDesign/methodology/approach: Informed by the ideational school of institutionalism, a qualitative documentary study was conducted to analyse the policy addresses, speeches, and blog posts delivered by the then CE, which were all available on the website of the HKSAR Government. It was through a thematic analysis of the database that themes and sub-themes were generated for the discussion. Representative verbatim quotes are used for illustrating some of the youth policy ideas and discourses promoted under the Leung's regime.\nFindings: The findings suggest that the governing youth narratives could be categorised into two interrelated themes: behavioural patriotism and economic opportunism. The notions of youth development constructed in the narratives of Leung shape the definition of youth-related problems and solutions in relation to national identity and global competition.\nResearch limitations/implications: This study focusses on the previous term of HKSAR Government that cannot fully reveal the extent of policy continuities and changes. Yet, it could outline the overall picture and address the shortcomings of Hong Kong's current youth policies viewed from both normative and cognitive perspectives. Methodological and analytical implications can be drawn for further studies on policy ideas and discourses.\nOriginality/value: The paper has two major contributions; the first of which is the illustration of an analytical framework connecting contents, tools, and justification of policies for capturing the dynamics and complexities of youth policies. Second, the findings of this study develop a critical understanding of neoliberal youth policies in both economic and moral senses that pose new challenges to young people and policy makers.\n",
"role": "assistant"
}
] |
Factors inhibiting the internet adoption by base of the pyramid in India
|
[
"Digital divide",
"Internet adoption",
"Bottom of Pyramid (BoP)"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: In almost every country in the world, the use of the internet services has grown significantly during the past decade. The internet facilitates the interaction between individuals and businesses to share information, resources and services (Sahin, 2010). Millions of individuals today are connected to each other via the internet (Almarabeh et al., 2016). India is the second largest online market and is ranked only behind China. Despite such a huge base of the internet users in India, only 26 per cent of the Indian population had access to the internet in 2015 (www.statista.com/topics/2157/internet-usage-in-india/).\n2. Digital inequality among the base of pyramid segment: The degree, to which people are able to gain benefit of available information by using the internet, depends on their capability to navigate the online environment. Those who are unable to involve with the technologies are at the disadvantage of realising the benefits that the online environment may offer (McCloud et al., 2016). For example, e-mail allows the instantaneous and asynchronous communication that eliminates the communication barriers present in, in-person local, and telephone contact. Other internet technologies, such as a website, can lessen the cost of local participation (Hampton, 2010).\n3. Research methodology: Triangulation, specifically with the usage of qualitative and quantitative (QUAL/QUAN) sequence, can help in performing an in-depth analysis of a particular phenomenon (Creswell, 1998; Tashakkori and Teddlie, 2009). Hence, we have used a mixed-method study with QUAL/QUAN sequence in this research. The mixed-method study is a useful instrument for the information systems researchers for a number of reasons. First, this approach can tackle the exploratory as well as confirmatory research questions simultaneously. Second, it can compensate the weaknesses of a single method. Third, it provides the prospects of obtaining several divergent views (Tashakkori and Teddlie, 2009; Venkatesh et al., 2013).\n4. Findings of the qualitative study and hypotheses development: Following is the discussion on five themes identified as potential inhibitors to internet adoption by the BoP segment in India:\n5. Discussion and policy implications: Digital inequality is one of the most prominent issues in the knowledge economy (Hsieh et al., 2008). Literature review shows that sometimes poor people are unable to gain the benefits due to their lack of capability to use the technologies (McCloud et al., 2016). Knowledge of the inhibitors to adopt the internet by the BoP segment can be useful to bridge this gap. Thus, there is a need to discover these inhibitors, which is the focus of this research. The results of this research confirm that the factors such as lack of digital literacy, complexity and language barrier play a significant role in inhibiting the internet adoption by the BoP segment. However, there is no impact of perceived risk and cost on the internet adoption.\n6. Conclusion and further research: This research addresses the factors inhibiting the internet adoption by the BoP segment in India. The proposed model provides the knowledge regarding different factors negatively affecting the intention to use the internet by the BoP segment. Initially, a qualitative data analysis of interviews with ten BoP people revealed potential inhibitors. These factors were hypothesized to inhibit the internet use. Later, the survey data were collected from 194 BoP people to test the hypotheses empirically. We found strong support for some of the hypotheses in the proposed research model. This information may be used to leverage the use of the internet by BoP people. The study also explains how this information can be used to increase the use of the internet among the BoP segment. However, this paper only focuses on the BoP segment in India. Therefore, we suggest that future researchers can analyse how India is different to other developing countries. Furthermore, this research finds that there is no impact of perceived risk and cost on the internet adoption by the BoP segment in India. Future researchers can explore the reasons for these findings.\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to explore the factors inhibiting the internet adoption among base of pyramid (BoP) and to develop further insights of such factors.\nDesign/methodology/approach: Researchers used a mixed-method study with QUAL/QUAN sequence in this research. Given the dearth of research on the internet usage by the BoP segment in India, the authors began with the exploratory qualitative study. Semi-structured interviews in local language were carried out with the people belonging to the BoP segment in India using common interview protocol. Face-to-face interviews were conducted for 20-30 min with ten people. Thematic analysis (Braun and Clarke, 2006) was conducted on the interview transcripts which lead to five themes. These themes were identified as the potential inhibitors to the internet adoption by the BoP segment in India. This step was further followed by a confirmatory quantitative study.\nFindings: The results confirm that the factors such as lack of digital literacy, complexity and language barrier play a significant role in inhibiting the internet adoption by the BoP segment. However, there is no impact of perceived risk and cost on the internet adoption. The significant relationship between lack of digital literacy and intention to use implies that if the BoP segment is capable to locate, evaluate and use digital information, it is more likely to use the internet. Similarly, the relationship between complexity and the intention to use suggests that if the BoP segment finds it easy to use and understand the internet, it is more likely to use it.\nPractical implications: This research has the implications for government departments and policymakers that are responsible for promoting use of the internet. Therefore, such departments and policymakers are advised to channelize their effort on resting the digital literacy, especially in terms of internet adoption. Findings of this study show that digital literacy, complexity and language barriers are major inhabitants in the internet adoption. Directed and focused government policies and initiatives such as subsidized seminars and training programmes specially designed for the BoP segment can help in increasing internet adoption.\nSocial implications: The adoption and use of the internet services by BoP have major implications for digital equality. Therefore, it would be important to explore the factors inhibiting the internet adoption among BoP.\nOriginality/value: This paper not only explores the factors inhibiting internet adoption among BoP in developing countries such as India but also provides deeper insights to these factors by collecting first-hand information from target segment. The finding of this research provides meaningful inputs to policymakers and also to industry to remove digital divide in the target population. This is how this research adds value to the existing knowledge available in this domain and it also provides agenda for future research.\n",
"role": "assistant"
}
] |
Corporate human rights social responsibility and employee job outcomes in Kenya
|
[
"CSR",
"Africa",
"Human rights",
"Kenya",
"Employee retention",
"Employee satisfaction",
"Corporate human rights social responsibility",
"Flower exporters",
"Labor rights"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Organization studies on human rights (HR) and corporate social responsibility (CSR) have been fragmented and asymmetrical. In retrospect, few studies on HR have been linked to CSR and vice versa. Accordingly, studies on HR have been peripheral to organization studies, and thus setting HR and the corporate organization worlds apart (Lozano and Prandi, 2005). Whilst, concern for HR in Africa has been on the rise, this has not been commensurate with rise in scholarly work. Until very recently, HR has played a rather minimal and peripheral role in the conceptualization of CSR (Wettstein, 2009). Consequently, CSR has been prominently perceived as internal, and HR as external to the organization.\nCorporate social responsibility and human rights: While some argue that CSR and HR have different meanings, others argue that indeed the two concepts are similar. Thus, CSR is a fuzzy concept with unclear boundaries and debatable legitimacy (Lantos, 2001). And also a transversal issue that impacts the organization in many different ways. Variously described as an oxymoron, arising from its duplicity and multiple initiatives, CSR remains an embryonic and contestable concept (Winsor, 2006). However, the meaning and value of CSR may differ in various contexts, depending on local factors including culture, environmental conditions and legal framework (Galbreath, 2006). Accordingly, different cultures will emphasize different values (Burton et al., 2000) leading to lack of universal CSR and HR position in Africa. CSR represents the commitment by firms to behave reasonably and responsibly and contribute to local economic development while improving quality of life of its employees and the local community (Cheruiyot, 2009). More significantly, CSR is essentially about companies combining economic, social, environmental and HR interests for the good of the company and its stakeholders (European Commission, 2001).\nFrom CSR and HR to CHRSR: Relationship between CSR and HR is not properly understood with some scholars characterizing them as diametrically unique. While CSR approach tends to be top-down and voluntary, HR is bottom-up, moral and legalistic, with the individual at the centre, not the corporation (Avery, 2006). HR is based on the inherent dignity of every person. They are those basic rights and freedoms to which all humans are entitled. Companies are obliged to respect HR at all times, not just when it suits them. It is rooted in law. Respecting and protecting them is not a matter of choice. It should be part of a mainstream strategy not just as part of its CSR strategy. On the other hand, firms not adopting a HR framework could be fundamentally exposing themselves to risk.\nEmployee job outcomes: Whilst some literature focus on labor rights as antecedent to employee job outcomes (Cheruiyot and Maru, 2012) others allude to impact of organizational justice on employee satisfaction and commitment (Ishak and Alam, 2009; Folger and Cropanzano, 1998). The fragmentation in the literature and its diffuse nature could be as a result of among others its multidimensionality. This is particularly critical in African unique work environment. We therefore propose that:\nLink between CHRSR and employee job outcomes: Many scholars conceive CSR as encompassing internal and external dimensions (Jamali et al., 2008). Internally, CSR concerns according due diligence to internal stakeholders by addressing issues such as skills and education, workplace safety, working conditions, human and labor rights, equity considerations, equal opportunity and health and safety (Jones et al., 2005). External CSR focus on diligence to their external economic, social stakeholders and the natural environment (Munilla and Miles, 2005). This growing body of literature suggests that a firm's involvement in CSR efforts influence stakeholders' generalized reaction to the organization. However, relatively neglected in the literature are current and prospective employees as a critical stakeholder group. Inside the firm, attention to CSR could enhance employees' response to the work environment.\nStudy methodology: Research design\nStudy results: This section provides descriptive statistics, factor analyses of CHRSR and finally multiple regression analysis on the effect of CHRSR on employee job outcomes.\nDiscussion, conclusions and implications: This article explored the dimensionality of CHRSR and employee job outcomes in a Kenyan context among three leading flower-exporting firms. Further, it examined the link between the two constructs. Apart from establishing the multidimensionality of the constructs, findings also showed a clear link between CHRSR and employee job outcomes.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to identify dimensionality and evaluate effect of corporate human rights social responsibility (CHRSR) on employee job outcomes among three leading Kenya's horticultural exporters based in Naivasha, Kenya.\nDesign/methodology/approach: Exploratory and explanatory survey was utilized on a sample of 450 employees selected using a multistage sampling technique. Structured questionnaire was used to extract both nominal and ordinal data, the latter's items were anchored on a five-point Likert scale. Descriptive, principal component analysis (PCA) and multiple regression was used to analyze the data.\nFindings: PCA of 32 items representing CHRSR yielded ten components while that of employee job outcomes yielded seven components reflecting the multidimensional nature of the constructs. Results showed that the selected firms have largely satisfied both HR and moral expectations. Regression results showed generally significant effect of CHRSR on employee job outcomes such as employee retention, satisfaction, commitment and pride.\nResearch limitations/implications: A business case for CHRSR exists. This implies that compliance with the law, adoption of HR-related corporate social responsibility (CSR) is critical in enhancing employee job outcomes and overall organization performance.\nOriginality/value: The study integrates CSR and HR into CHRSR. It demonstrates its multidimensional nature and utilizes statistical analysis to explore its relationship with employee job outcomes in an African export sector context and argues that CSRHR is more critical in Africa than more regulated contexts.\n",
"role": "assistant"
}
] |
In the eye of the beholder: How self-construal influences service evaluations following observations of others' service experiences
|
[
"Service failure",
"Service recovery",
"Self-construal",
"Justice",
"Observing customers"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Service failure costs firms billions of dollars every year. According to estimates, in 2016, in the USA alone, businesses lost close to $62bn because of poor customer service (Hyken, 2016). Understanding how customers respond to service failure and recovery efforts, therefore, is a crucial part of firms' efforts to ensure customer retention and satisfaction (Smith et al., 1999).\nTheoretical framework: The impact of service failure and recovery on observing customers\nStudy 1: This study was designed as a first test of the hypothesis that self-construal (independent vs interdependent) moderates the effect of service failure and levels of compensation on observers' satisfaction with a service encounter. We expected that following the observed service failure of another customer, participants high in interdependent self-construal would report lower satisfaction relative to those high in independent self-construal. Further, we predicted that the effect of compensation level would be stronger on the former than on the latter.\nStudy 2: This study was designed with several objectives in mind. First, we wanted to test the robustness of the hypothesized effect of self-construal on service satisfaction using a different setting (i.e. a hotel scenario) and a different operationalization of self-construal, i.e. as a stable tendency to view the self as an autonomous individual versus as part of a larger social collective. Second, we wanted to obtain evidence for the mechanism driving this effect. We had hypothesized that individuals' varying sensitivity to others' treatment is mediated by perceptions of justice. In particular, we expected that following service failure, interdependent (relative to independent) participants would report perceptions of lower interactional and procedural justice. We expected that interdependent participants would report perceptions of higher distributive justice when the target received high (vs low) compensation. We expected these differences in perceived justice to mediate the effect of self-construal on satisfaction.\nStudy 3: Study 3 was designed with several objectives in mind. First, we wanted to ensure that the hypothesized effects of observing others' experience on satisfaction hold even when the observers' own experience with that service is satisfactory. For this purpose, we incorporated information about the observers' own service experience in the scenarios. Second, we wanted to enhance the practical implications of our research using a managerially relevant self-construal prime. Specifically, we tested whether service providers can influence customers' salient self-construal (and thus their perceptions of the service) by varying elements of the service context (e.g. promotional materials) that are under the control of the provider. We further highlighted the practical implications of the research by examining the effect of self-construal and type of experience on observers' re-patronage intentions. Previously, studies have demonstrated a close link between satisfaction and perceptions of fairness and re-patronage intentions (Anton et al., 2007; Mattila et al., 2014). Thus, we expect that the impact of self-construal on responses to failure and recovery will be captured by behavioral measures such as intentions to return to the service provider. Including these additional measures would provide converging evidence for the robustness of the hypothesized effects of self-construal.\nGeneral discussions: Many services are delivered in public settings, i.e. in the presence of multiple customers. In such settings, for every customer experiencing service failure, there may be many more who witness the incident and could be indirectly affected by it. The downstream effects of service failure are further magnified because of the growing use of online forums for sharing consumer experiences. Thus, understanding how service failure and recovery efforts impact observing, as well as target, customers becomes increasingly relevant. This study investigates how the self-construal of observing customers - the extent to which they see the self in relation to others - influences their reactions to others' service experience.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to investigate the effect of interdependent vs independent self-construal on service satisfaction, following the observation of failure and recovery experienced by a fellow customer.\nDesign/methodology/approach: Three experiments were conducted to test the research hypotheses.\nFindings: After service failure, interdependent observers react less favorably compared to independent observers. After high recovery compensation, interdependent observers react more favorably compared to independent observers. The effects are driven by differences in perceived interactional and distributive justice.\nResearch limitations/implications: This study uses three scenario-based experiments to test the hypotheses. While providing greater control of the experimental conditions, the external validity of the results is to some extent sacrificed. Moreover, this research does not investigate observers' reactions to the interactional aspect of recovery.\nPractical implications: When handling service failure, firms are required to anticipate and address not only the responses of the target customers involved but also those of potential observers. Providers can accordingly use available customer information to gauge customers' likely self-construal and to adjust their service delivery and recovery tactics. Providers can influence observing customers' reactions by creating a servicescape that activates a desired self-construal.\nOriginality/value: This research is one of a few to examine the effect of service failure and recovery on observing customers, and the first to do so via the lens of self-construal. It contributes to the literature on service failure and recovery and the literature on self-construal and has practical implications for service providers. The value of this research is further highlighted given the increasingly public nature of services and the multicultural context of service delivery.\n",
"role": "assistant"
}
] |
Brand authenticity: model development and empirical testing
|
[
"Brand identity",
"Brand trust",
"Authentic brand",
"Brand authenticity"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nAn executive summary for managers and executive readers can be found at the end of this issue.: The constant growth of the number of brands, similar product offerings and fast imitation of product innovation have led to high levels of brand parity (BBDO Consulting GmbH, 2009) and ever-increasing brand competition. In response, companies have significantly increased their communication efforts (Kotler and Keller, 2011). However, the cognitive processing capacities of consumers are limited, and the current high levels of marketing communication have been shown to lead to consumer reactance (Holt, 2002).\nThe authenticity concept: In reference to objects, the authentic is often considered as the \"original\" and is contrasted with the copy (Peterson, 2005). For example, an original Picasso artwork is authentic, as it was created by the artist himself. This type of authenticity is objective authenticity, i.e. the authenticity is inherent in the object itself.\nAntecedents of authenticity: Applied to the antecedents of authenticity, attribution theory is an umbrella term for various attempts to explain personal authenticity by examining how individuals determine the causes of behavior and events.\nEffects of authenticity: Perceived authenticity has been argued to increase the level of trust. Trust itself is a risk-oriented construct, which is important in situations where the trusting party is unsure of future actions of the trust-taker (Delgado-Ballester et al., 2003; Sirdeshmukh et al., 2002). The perceived risk results out of the uncertainty that the trust-taker will fulfill a task in the interest of the trusting party (Rotter, 1980; Morgan and Hunt, 1994). A theoretical explanation for the hypothesized positive influence of authenticity on trust is that future actions of authentic people are likely to be consistent with their past behavior and hence enjoy considerably higher predictability. As their behavior is primarily led by their identity, authentic people follow guiding principles. Hence, relationships with authentic people are therefore more manageable (Chan et al., 2005). Furthermore, competence and integrity of the trust-taker which have been proven in the past can be projected into the future. This in turn decreases the level of perceived risk and increases the level of trust.\nThe concept of brand authenticity: The identity-based brand management model comprises two main components, brand identity and brand image (De Chernatony et al., 2011). Brand identity is defined as \"those sustainable cross-spatiotemporal attributes of a brand which determine the brand's essence from the perspective of the internal target group\", e.g. employees, executives and intermediaries (Meffert et al., 2012, p. 361). In contrast, the brand image represents the exterior view of a brand. It is understood as \"condense and judgmental perceptions about a brand fixed in the psyche of relevant external target groups\", e.g. customers and consumers (Meffert et al., 2012, p. 364). The process of formulating the brand positioning must take into account the needs and expectation of consumers (De Chernatony et al., 2011). This market-orientation implies that companies run the risk of communicating brand attributes which are desired by the external target groups, but which do not reflect the true identity of the brand. Here brand authenticity can enhance the identity-based brand management model.\nCausal model of brand authenticity: The conceptualization of authenticity presented earlier can be translated into a causal model in which individuality, consistency and continuity of a brand represent the antecedents of brand authenticity. Within this study, \"individuality\" is defined as the unique way in which the brand fulfills its promise. \"Consistency\" measures those attributes which are expressed by its reiterative current brand behavior. Hence, high consistency is perceived if the brand promise is fulfilled at every single brand touch-point. \"Continuity\" refers to core brand attributes remaining stable over a longer period. Hence, the consumer judges continuity to be high if the present brand promise reflects past brand behavior. This study hypothesizes therefore (Figure 1):\nResearch methodology: The operationalization of the model is detailed in Table AI: three items were used to measure the individuality of a brand (Table AIII). These items were adapted from the uniqueness scale developed by Netemeyer et al. (2004). The items for measuring \"brand-consistency\" and \"brand-continuity\" originate in Erdem's and Swait's work (Erdem and Swait, 1998). Some reformulations were made to adapt the items closer to the scope of this study. These changes ensure that respondents easily realized the difference between the present-oriented construct of \"brand-consistency\" and the past-oriented construct of \"brand-continuity\". The resulting scales consist of three indicators each (Table AI and Table AII). The items for the brand trust measurement model were taken from the developed and already empirically tested trust scale of Chaudhuri and Holbrook (2001). However, one of their four items was excluded from the study, namely, the indicator: \"brand [X] is an honest brand\". As the authors share Morgan and Hunt's (2004) view that \"honesty\" is an antecedent of brand trust, this item could not be included in the scale for reasons of validity (Morgan and Hunt, 1994). The final brand trust measurement model hence comprises three items which contain statements about how safe and comfortable people feel when they rely on the fulfillment of the brand promise (Table AIV).\nEmpirical results: evaluation of the measurement models: All constructs of the research model are reflexive measurement models, i.e. the indicators are observable consequences of the construct. To ensure the reliability and validity of the measures, they were analyzed using the quality criteria proposed by Hair et al. (2011). Initially, the brand authenticity construct was tested for its indicator \"reliability\". The indicator \"reliability\" is fulfilled if each standardized indicator loading is higher than 0.70. The two reverse-coded indicators of the brand authenticity scale were both below a score of 0.47. One explanation for the weak loadings might be that the reverse-coding caused comprehension problems. To improve the quality of the scale, these items were eliminated. Indicator reliability was then re-tested for the revised scale. The loadings of the remaining four indicators were all above 0.70 (Table III). The assessment of the construct \"reliability\" in PLS focuses on composite reliability. This is comparable to Cronbach's alpha and serves as an estimate of a construct's internal consistency. With a value of over 0.90, the threshold of 0.70 was surpassed (Table III). In PLS, the validity assessment of the reflective measurement model focuses on convergent and discriminant validity. For convergent validity the average variance extracted (AVE) is examined and should be 0.5 or higher. With an actual value of over 0.71, this criterion is met. For the assessment of discriminant validity, the Fornell - Larcker criterion (Fornell and Larcker, 1981) was applied. As the AVE of the brand authenticity construct is higher than the latent construct's highest squared correlation with any other latent construct of the structural model, this criterion is well-matched (the result of these analyses are not presented here in the interest of brevity). The remaining analyses of brand individuality, brand consistency, brand continuity and brand trust are all significant and presented in Table III.\nEmpirical results: evaluation of the structural models: The primary evaluation criteria for a structural model are the R2 measures of the endogenous constructs and the level of significance of the path coefficients. In marketing research studies, R2 values of 0.66, 0.33 and 0.19 are described as substantial, moderate and weak, respectively (Chin, 1998). For the tested models, the explained variance of brand authenticity ranges from 63 per cent to 70 per cent (R2Overall = 0.65; R2Beer = 0.70; R2Fast-food = 0.63), which can therefore be classified as \"substantial\". Moreover, all path coefficients of the exogenous variables are positive and significant (Figure 2). Therefore, the sample showed that the perceived individuality, consistency and continuity of a brand are positively related to its perceived authenticity.\nConclusion and scope for future research: The positive and strong relationship between brand authenticity and brand trust found in this study provide support that authenticity can be regarded as an important success factor for branding. Furthermore, a positive causal relationship between brand individuality, brand consistency and brand continuity on brand authenticity was empirically confirmed. Therefore, fulfilling the brand promise in an individual, consistent and continuous way can be regarded as vital for building brand authenticity. Individuality of a brand was shown to be the least important factor, while consistency and continuity proved to be of major importance for building brand authenticity.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to develop a model of brand authenticity and analyze the antecedents and effects of the construct. Although there is no doubt about the relevance of authenticity in personal relationships, published research has yet not thoroughly explored the concept's meaning in reference to brands.\nDesign/methodology/approach: Based on socio-psychological attribution theories and grounding on the identity-based brand management approach, a causal model of brand authenticity is developed. The hypothesized relationships are analyzed using the partial-least-squares approach. The primary data are based on an online survey conducted in Germany (n = 600). The respondents were asked about fast-food and beer brands.\nFindings: The data show that brand authenticity positively impacts on brand trust. Furthermore, the key antecedents in the model (consistency, continuity and individuality of a brand) drive the perception of brand authenticity as hypothesized.\nResearch limitations/implications: The model should be tested in further product categories and moderators should be integrated.\nOriginality/value: The findings suggest that authenticity is perceived when a brand is consistent, continuous and individual in its behavior. Nevertheless, the empirical results indicate that the factor individuality has the lowest influence on perceived brand authenticity. This is an interesting finding, as being \"unique\" is commonly regarded as an important success factor in branding. Although the study's findings confirm its relevancy, they relativize its importance: being consistent, meaning that a brand fulfills its brand promise at every brand-touch point, and being continuous, meaning that the brand promise reflects the essential core of the brand, are of major importance.\n",
"role": "assistant"
}
] |
Addressing dynamic issues in information security management
|
[
"Data security",
"Information systems",
"Generation and dissemination of information"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Progress in information technology is currently the main engine for productivity gains in society (Hagsten, 2009; Brynjolfsson and Yang, 1996). Society also depends on and requires safe and secure IT infrastructures. The very engine that drives the progress in IT infrastructure simultaneously creates new threats to the security of the IT infrastructure. These threats cannot be predetermined and are uncertain in their revelation. Organizations face various security issues due to this uncertainty rooted in technology, obsolescence of their security measures, needs for changes in security requirements and calls for re-evaluation of IT products. These together make IT an uncertain infrastructure. Large organizational and societal level infrastructures require long and costly IT security evaluations based on Common Criteria (CC Portal, 2009b). Also, these infrastructures are the most obvious targets of new security threats and therefore require appropriate protection. Long security evaluation processes are a fundamental problem because by the time the evaluation is done, new threats have emerged, making the evaluation obsolete. Still, organizations are required to deal with these issues at the organizational level in their own environments in which they have the possibilities to adjust changing requirements to dynamically emerging problems and issues. In this research, these aspects have in the past been addressed as three separate research problems:1. how to deal with dynamic security requirements (Abbas et al., 2009c);2. how to deal with externalities (Abbas et al., 2010); and3. how to rapidly re-evaluate an existing IT infrastructure against constantly changing security threat scenarios (Abbas et al., 2009b).Options-based method from corporate finance domain have in this research been adapted and used in IT security requirements analysis and evaluation. In this paper, a coherent methodology is presented around three related problems. Each of these problems is given a more in-depth treatment and substantiate the arguments by applying the proposed methods to a societal-level information-exchange mechanism used by Swedish civil authorities, the SHS (Helenelund et al., 2007). At present, this method is well structured but manually applied. However, the ultimate objective of this research is to elaborate on how this could be automated in the future.The paper is organized as follows: as the methodology depends on real-options theory, it begins by introducing this theory and its salient features in Section 2. Section 3 elaborates on adapting options theory features and methodology building to target IT security issues. Sections 4-6 present an in-depth review of the state of the art and address each problem individually. Section 7 presents conclusion and future intentions of this research.\n2 Options theory advantages and related work: Real options analysis (ROA) is primarily a corporate finance method that provides strategic guidelines to reduce risk and manage uncertainty for investment decision making. The underlying concept of ROA is the use of real options that considers an investment decision as the \"right\" not the \"obligation\" and provides flexibility to make, abandon, expand or shrink the capital investment (Harvey, 1999). It provides mechanism for exploration of various alternative actions and chance events in terms of time are analyzed using option tree. Chance event probabilities are assigned based on historical data and modeling or judgments by experts (Chatterjee and Ramesh, 1999). The technique emphasizes staging the investment, considers alternative solutions and provides the ability to alter decisions based on the outcome/uncertainty revelation with the time. For example, during the lifetime of a product development, a firm has more than one strategic option that is; the firm may suspend production if the production cost exceeds the product value and can later be resumed when the conditions become favorable. These strategic options are the real options that are normally ignored in standard valuation techniques like net present value (NPV) and return on investment (Faboozi, 2007). Information security management issues (changing requirements, re-evaluation needs and externalities) caused by technological uncertainty have the same problem of being dealt rigorously with available solutions while ignoring the hidden information with uncertainty revelation. Several methods for predictive analysis like architecture trade-off assessment method (ATAM) (Byrnes and Kyratzoglou, 2008), attributes-based architecture styles (ABAS) (Kazman and Klein, 2000) and architecture-based analysis for complex systems (ABACUS) (Dunsire et al., 2005) are well known in this area. Maxwell et al. (2008) and Parakhine et al. (2008) also present predictive techniques for non-functional requirements. The methodology presented in this paper has the significance of capturing value for decision making during uncertainty that may reveal useful strategic and long-term information which is normally ignored in other predictive techniques (such as ATAM, ABAS and ABACUS). To cope with uncertainty issues, the methodology provides:* the ability of mid-course decision management;* efficient mechanism for exploration of alternative solutions; and* flexibility in change-management process.It considers multiple architecture evolutions for alternative solutions using different security metrics (performance, investments, etc.) to analyze and predict for each branch deterministically with good confidence.Options theory has numerous success stories in science and technology, oil and gas, R&D and IT security investment projects (Tansey and Stroulia, 2007; Chatterjee and Ramesh, 1999; Li and Su, 2007; Gordon et al., 2003). The core concept of options valuation technique has also been deployed in various diverse domains following different work processes in software engineering for example, where Ipek Ozkaya applied real-options theory to software architecture design decision making (Ozkaya et al., 2007). The work presented in this paper emphasizes deducing the core theme/notion of options theory and its deployment in a different perspective to target uncertainty problems in information security management. The research focuses on strategic security decision-making aspects during uncertainty as presented in Sections 4-6. But before going into details; it will be worth mentioning that how real options-based concepts are made coherent to IT security work processes using options-based security-oriented framework. Section 3 presents a brief description of this framework.\n3 Security-oriented real options-based methodology: ROA as a corporate finance method relies on Black Scholes, binomial lattices, Monte Carlo simulation to calculate monetary values and is more inclined towards mathematical models and expressions (Mun, 2005). This infrastructure provided by options theory could not be as such applied to IT security issues and work processes (that follow system level analytical methods). Therefore, an adaptability framework was devised (Abbas et al., 2009a) to transform/bridge options theory concepts to address IT security issues and described in Table I in a nutshell. The objective was to study/analyze options theory problem solving mechanism and substantiate the concepts in IT security perspective.In Table I, the first column describes phases and features of ROA, whereas the second column describes the key concerns of IT security management with respect to security requirement specification, product development and security evaluation. The ROA core concepts and work processes were analyzed for applicability and substantiated with each IT security uncertainty problem considering security management work processes in an organization. The inference provided basic rules of thumb and guidelines from the perspective of the security-oriented methodology. The phases and salient features of this resulting approach (in the third column) are based on analyzing organizational structure (security objectives and policies), information security management phases/constraints, applicable ROA features and IT security problems nature.It is also worth mentioning here that the core feature of options theory is the use of various \"options\" to formulate strategy according to appropriate scenarios. These options were also streamlined according to security-oriented options-based methodology and are described in Table II.Thus, the adaptability framework resulted in a security-oriented real options-based methodology comprising of three main phases:1. identification;2. formulation and analysis in options theory context; and3. recording information.The detailed description and illustration of each uncertainty problem in IT security using this methodology is presented in Sections 4-6. Table II is referred to in each section for formulating solutions.\n4 Information security externalities: Externality is an economic concept for the effects borne by the party that is not directly involved in a transaction (Cornes and Sandler, 1996). Externalities could be positive or negative. A security system deployed in an organization may also cause externalities for other interacting partners/systems. For example, a strong authentication module provided by a bank's web site causes positive externality for the e-commerce system that redirects its purchases to the bank's web site for authentication of users. Negative externalities in IT security are uncertain and may appear in many forms, for example obstructing legitimate access, privacy violation (Cavoukian, 2009) and unavailability of services. It is common practice that organizations barely consider positive externalities although they can be beneficial; and ignore negative externalities that might cause vulnerabilities for themselves and their interacting partners/systems. The methodology presented below provides a cost-effective mechanism for organizations to internalize their uncertain externalities considering:* Liability concept from Schneier (2008) that infers \"everyone causing vulnerabilities should also be held responsible for their impacts to others\".* Lowest cost strategy by Coase (1960) that says externality burden should be placed where the cost is minimum.Externalities caused by an information security system are uncertain in their revelation and cannot be predetermined before security system's deployment. Therefore, the issue has been considered using options-based approach to provide flexibility for internalization process and consider all parameters (organizational constraints, budget and uncertainty revelation) in solution formulation. It has been demonstrated by the e-society (ESAM)[1] (ESAM) case study and has following three main phases:1. identification;2. formulation and analysis in options theory context; and3. recording information.Each phase may include sub-phases as described below.(I) Identification\n5 Dynamic security requirements management: Rapid technological development raises novel security concerns for an organization. Their existing security measures become obsolete and technology becomes ineffective as new vulnerabilities appear with progress in technology. Organizations strongly require organizing and managing dynamically changing requirements and keep their system up to date. Even the product whose security has been certified by (CC Portal, 2009b) requires a lengthy process to manage changed requirements to overcome the inability of static solutions to deal with the dynamic problems. The issue is addressed using options-based approach that helps to formulate organizational security requirement management strategy for a continuously changing environment. It provides strategic guidance for decision makers to effectively consider available and upcoming technologies, current and future organizational policies and budget plans. The methodology is illustrated using ESAM system for dynamic security requirement management process and comprises of following phases.(I) Identification\n6 Ongoing evaluation/re-evaluation of IT security services/products: The third problem addressed using options-based approach is the organizational need of continuous security evaluation for their IT products, services, methods and technology. The re-evaluation becomes inevitable in this dynamic environment when novel threats and vulnerabilities continuously appear with the progress in technology. Even evaluated/certified products need to be re-evaluated and are required to submit change information and regain certification. CC Portal (2009b), for example, is an established methodology for security evaluation of IT products and is recognized in various countries (CC Portal, 2009a). CC evaluated/certified products become obsolete as the technology progresses and new threats and vulnerabilities appear. Organizations require modifications in their existing security mechanisms and also submitting change summary to CC lab to regain certification (National Security Agency, 2008). As the technology changes at a rapid pace, the recertification becomes a difficult and cumbersome process. A continuous security evaluation mechanism within the organization is a critical need to achieve information security objectives. It was also inferred from the study of analyzing CC issues (Abbas et al., 2008). Therefore, an infrastructure is introduced to avoid lengthy re-evaluation processes and to manage re-evaluation at organizational level. It cannot be claimed as a replacement to CC but the purpose of this research is to provide a tool for decision makers for tackling the security re-evaluation needs of an organization that is:* propose alternative evaluation strategies for changing security environment, considering organizational polices and time constraints;* explore, build and prioritize test criteria for a certain security mechanism based on system state; and* expedite the evaluation process.It is worth mentioning here that the re-evaluation process is tied with dynamic security requirement management process (described in Section 5) to address changes during evaluation. The re-evaluation process is named as option-based re-evaluation (OBR) that is the modified version of options-based evaluation (OBE) (Abbas et al., 2009b, c). The main phases of OBR are as follows.(I) Identification\n7 Conclusion and future work: The paper introduced a novel methodology to deal with three specific information security-related problems:1. information security externalities;2. dynamic security requirements management; and3. ongoing evaluation/re-evaluation needs for IT security services/products.In this research, an options-based framework is created and adapted to the specific IT security issues such that they can be formulated and analyzed according to options theory. The resulting approach is capable of discovering value during uncertainty for investigating optimal solutions from a series of alternatives and providing strategic guidance for information security management issues in a dynamic environment. It was demonstrated using real cases - SHS and ESAM, where the option-based mechanism provided significant outcomes for formulating IT security management solutions under uncertainty. The paper presented evaluation of the basic concepts of this methodology through manual illustrations from SHS and ESAM system. The result of each of the three problems (elaborated in Sections 4-6) will be used in near future as specifications for an automated solution (Automated Uncertainty Management System for Information Security (AUMSIS)) for uncertainty management at organizational level. As an ultimate objective of this research, AUMSIS will serve as a decision-making tool, deployed in an organization for uncertainty issues management in IT security.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to address three main problems resulting from uncertainty in information security management: dynamically changing security requirements of an organization; externalities caused by a security system; and obsolete evaluation of security concerns.\nDesign/methodology/approach: - In order to address these critical concerns, a framework based on options reasoning borrowed from corporate finance is proposed and adapted to evaluation of security architecture and decision making for handling these issues at organizational level. The adaptation as a methodology is demonstrated by a large case study validating its efficacy.\nFindings: - The paper shows through three examples that it is possible to have a coherent methodology, building on options theory to deal with uncertainty issues in information security at an organizational level.\nPractical implications: - To validate the efficacy of the methodology proposed in this paper, it was applied to the Spridnings-och Hamtningssystem (SHS: dissemination and retrieval system) system. The paper introduces the methodology, presents its application to the SHS system in detail and compares it to the current practice.\nOriginality/value: - This research is relevant to information security management in organizations, particularly issues on changing requirements and evaluation in uncertain circumstances created by progress in technology.\n",
"role": "assistant"
}
] |
Healthcare performance turned into decision support
|
[
"Key performance indicators",
"Sickness absence",
"Performance measurement",
"Decision support",
"Health care",
"Performance management",
"Absenteeism",
"Health services"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The health sector is currently under heavy political pressure. In recent years, rounds of layoffs and decreasing budget limits have prevailed in Danish public hospitals due to the economic recession that started in 2008 (Drachmann, 2010). Hence, a decreasing amount of available resources will have to manage an increasing number of patients. At the same time, managers also have to do an increasing amount of administrative work alongside their daily operational tasks. These challenges haunt managers in hospitals, who therefore need proper information in order to maintain the correct treatment of patients and at the same time focus on maintaining a good working environment. Thus, it is of utmost importance that the remaining staff attend work as much as possible. How this is achieved could be through the retention of currently available human resources in the workplace. High employee attendance can be sought by directing the searchlight towards the reasons for employee absence and investigating how this topic can be conceptualised.In an effort to conceptualise employee absence, the term \"sickness absence\" is elaborated. The term \"sickness absence\" itself is regarded by the author to be imprecise. As an example, understanding the term \"sickness absence\" literally does not correctly cover absence due to motivational reasons. Therefore, the term \"employee absence\" is introduced in this paper instead, covering absence due to both lack of motivation and being physically incapable of attending work.Few attempts to conceptualise employee absence have been carried out to date; those that have been undertaken have focused on motivation (Nielsen, 2008; Kristensen et al., 2006; Kaiser, 1998) or from an economic point-of-view (Barmby et al., 2003). The most cited model based on individual motivation on a unit level is still that of Steers and Rhodes (1978).Employee absence remains a complex phenomenon that can be caused by numerous factors (Duijts et al., 2007). Asking employees in their natural working environment, having fewer employees to treat increasingly more patients will intuitively result in a poorer service level, increased waiting times and cancelled operations. Paradoxically, the direct opposite situation has proven to be the case in the short term. Shorter waiting times, lower mortality rates and sickness absence and higher patient satisfaction were obtained at Frederiksberg Hospital (Drachmann, 2011). This fact speaks of high complexity in coherence with sickness absence.Common to most research is that a single or few factors is/are analysed in terms of sickness absence in a variety of sectors. These factors are, for example, the influence of work time (Laaksonen et al., 2010), workload (Roelen et al., 2007) and whether the employee feels that he or she is part of a social community (Bockerman and Ilmakunnas, 2008). The term \"social capital\", comprised of the elements \"reliance\", \"fairness\", and \"co-operation\", has been verified to impact on working conditions and employees' wellbeing and is receiving increasing attention from all kinds of organisations (Kristensen et al., 2008b). Since high scores in social capital are equivalent to better working conditions and higher employee contentment, it is easy to hypothesise that social capital may also influence employee absence patterns.Emphasis on a more holistic and integrative approach in absence management is suggested and desired (Dibben et al., 2001). In order to make such an approach, the determinants of employee absence need to be recognised, but unfortunately they remain elusive (Kaiser, 1998). Mapping and correlating formerly analysed factors with direct influence on sickness absence will be an important step to understand which determinants underlie sickness absence. When having pinpointed the determinants, the stepping stone for future interventional studies in evaluating the effect of given initiatives is prepared.For a more comprehensive assessment it is recommended to identify leading (determining future events) and lagging measures (the outcomes) (Evans, 2004). To identify the determinants of employee absence, many researchers carry out comprehensive observational studies involving broad pools of participants (Laaksonen et al., 2010; Nielsen, 2008; Michie and Williams, 2003; Munch-Hansen et al., 2008). In this study, the approach has been to extract information using only the scores from past employee satisfaction surveys.The aim of this paper is to provide a clear and tangible overview of the factors directly influencing employee absence. This is done by dividing the factors as objectively as possible into two subcategories, i.e. voluntary and involuntary absence (Chadwick-Jones et al., 1982). Identification of the factors leading to employee absence enables analysis of determinants through:* calculating Pearson correlations (Johnson, 2005); and* calculating social capital.Pearson correlations investigate the possible coherence of singled out factors with absence rates, while social capital involves a different and more conceptual approach. Encircling plausible determinants is thus performed using the two approaches. The data material used for testing the literature findings on the case material is employee satisfaction survey scores from a selection of Danish public hospitals. Interestingly, one of the case organisations exhibits very low absence rates, while another is burdened by high absence rates. Common to both is that they cannot explain why the absence rates are as they are. This fact suggests a decisive difference in one or more measurements in determinants. A practical framework showing the degree of contentment in determinants controlling absence rate in the public healthcare sector is proposed.\nMethodology: In the research community of the Operational Management Department at the Technical University of Denmark, the authors have previous experience with IDEF0, which originally was a technique applied to modelling manufacturing processes, but one that can be used for multiple purposes (Staccini et al., 2004). In this study the aspect of layer representation had previously been applied to analyse the linkages between patient and employee satisfaction. Employee absence was, among many other factors, identified as an important factor connecting the two types of satisfaction. This allowed deeper analysis of the contents contained in sickness absence, thus advancing one level in IDEF0 terminology. Such an overview serves as an initial comprehension of the complexity of employee absence while systematising the literature findings in respective categories.Analysing the causes of sickness absence\nResults: The derived results comprise two different approaches. First, the factors identified are correlated in turn with absence rates. Second, the social capital is calculated for selected departments.Links identified in the literature correlated with gathered case material\nDiscussion: There are many potential possibilities in performance measurement using questionnaires. Through the use of questionnaires applied at two healthcare departments, this study suggests a total of four overall elements to monitor when managing the complex phenomenon of reasons for employee absence.Concerning social capital, the scores from the four different questions are in correspondence with the previously published absence rates (Kristensen et al., 2008a). Only the scores from one of the hospitals differ slightly from the overall tendency, being an opposite development between social capital and absence level. The reason for this difference can be a variation in the formulation of questions used as indicators for the scores.At ATP the long-term focus on social capital has paid off in terms of high scores in employee satisfaction with the benefit of low absence rates (Kristensen et al., 2008b). The low absence rate at ATP must be interpreted by bearing in mind that employees have the possibility of working at home. Therefore, the customer centre has been the object of analysis, since the call centre requires employees to be present.Questionnaires are regarded as a potential method to investigate unwieldy issues such as possible reasons for employee absence (Traberg, 2010). Extraction of information through the employee satisfaction survey scores obtained can prove accurate, but there can be no guarantee that the management framework suggested in this study is applicable as a stand-alone tool. To support future findings, it is recommended that in-depth knowledge is gained about culture in a hospital department by applying a combination of social science and semiotics (Latour, 2005). The reason for recommending semiotics is due to its pragmatic approach comprehending three terms, i.e. humans, structure and technology. Although semiotics is a non-scientific approach, uniting these terms allows for in depth insight into the culture prevailing at a Danish public hospital. This knowledge can become explicit through interviews, workshops, and observations on site.The limitations of questionnaires are described in much of the research literature, focusing amongst other things on representative sample size (Skalland, 2011), interpretation and understanding of the questions asked (Johns, 1994) and confidentiality issues in the health sector. Furthermore, the responses collected may prove biased by the mood of the respondent. Employees showing a high degree of satisfaction are anticipated to reply positively throughout a satisfaction survey. The opposite result will be the case for employees with a low degree of satisfaction.However, the questionnaires analysed were approved in terms of their validity by external consultancy agencies, which speaks in favour of their statistical significance.Since it is not obligatory for public hospitals to run employee satisfaction surveys, such surveys are carried out rather sporadically and only once a year. Higher quality data would allow for the use of multivariate statistical analysis, adding additional validation to the results obtained.This fact is an obstacle in terms of data analysis and comparisons across departments. In the search to determine true findings, several hospitals were therefore contacted and asked for the results of employee satisfaction surveys that they had undertaken previously.When doing comparisons of survey scores, it is important to interpret the results with caution because of differences in the structure of the surveys. The various consultancy companies providing the surveys present different standards and use a wide variety of formulations in questions and presentations of the scores obtained between customers. In years to come, more data will be readily available to further test this study's suggested management framework for its precision.The results obtained from the interviews were basically gathered by applying the research strategy of case studies (Yin, 1989). One of the strengths of this particular strategy is that the researcher will gain in-depth knowledge about the subject of interest in its natural surroundings. Generalisation of the results must be done with care.Regarding future application of the suggested framework introduced in this paper, the factors identified through multiple case studies originate from the Danish healthcare sector. The application of this framework to other sectors has not yet been tested. It may be that there are other more accurate factors dominating the level of absence in the private sector.Moreover, one cannot neglect the variation between individuals making it difficult to pin down precise determinants in absenteeism that are valid for all. Generalisation of the results must be done with caution, since only a handful of implied parties have been included. To strive for a more solid basis for generalisation, the creation of a broader international case study that would involve a larger number of public hospitals is a way to obtain this. Having a broader span of data would allow for more in-depth statistical analysis, thus accepting or rejecting the current indicators for absenteeism.The framework is ready for use and can be merged with already existing questionnaires. Monitoring the risk of possibly absent employees will open up the possibility of taking preventive action if necessary.\nConclusion: Comprehension and specification of the determinants that control employee absence rates was the main focus of this study. The approach started by conducting interviews and scouting literature and afterwards analysing the scores from a wide selection of employee satisfaction surveys carried out in the public healthcare sector. Having performed the empirical study, provision of a tangible overview of factors directly influencing absence rates, analogous to an IDEF0 chart, was made. The links identified were correlated with concurrent absence rates to test for possible coherence in development over time.The final product of the study was a management framework which can provide healthcare managers with valuable and holistic information about the current status of the determinants that control employee absence rates. The management framework provides a practical and straightforward approach that transforms the examined theory into a practical solution, which represents the contribution of this study to sickness absence research. Additionally, the framework possesses a beneficial merging with satisfaction survey standards as well as an easily comprehensible and visual representation of scores. The manager is not only provided with insight into which factors generally contribute to absence rates, but can also quantify the effect of corrective actions. To support future findings from the use of the management framework, initiatives to improve employee satisfaction must be suggested on the basis of interviews and workshops with the employees concerned.Since the suggested framework is not static, it is an iterative process to evaluate the current determinants to test their validity.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this study is to first create an overview of relevant factors directly influencing employee absence in the healthcare sector. The overview is used to further investigate the factors identified using employee satisfaction survey scores exclusively. The result of the overall objective is a management framework that allows managers to gain insight into the current status of risk factors with high influence on employee absence levels.\nDesign/methodology/approach: - The research consists of a quantitative literature study supported by formal and semi-formal interviews conducted at the case organisations. Employee satisfaction surveys were applied to analyse the development over time of selected factors correlated with concurrent employee absence rates. Checking for causal results, comparisons with the included published literature findings were also carried out.\nFindings: - Four major clustered factors, three of which constitute the term \"social capital\", showed a high degree of connection with employee absence rates. The factors are general satisfaction, fairness, reliance and co-operation. Integrating the four elements in a management framework will provide valuable and holistic information about the determinants with regard to current levels of employee absence. The framework will be a valuable support for leaders with the authority to alter the determinants of employee absence.\nResearch limitations/implications: - Since a great part of the empirical material is supplied from the healthcare sector, the results obtained could be restricted to this sector. Inclusion of data from Arbejdsmarkedets Tillaegspension (ATP) showed no deviation from the results in the healthcare sector.\nPractical implications: - The product of the study is a decision support tool for leaders to cope with levels of employee absence. The framework is holistic and can prove to be a valuable tool to take a bearing of where to focus future initiatives.\nOriginality/value: - Gathering former observational studies in a complete overview embracing many relevant factors that influence sickness absence has not yet been attempted. Hospital management is provided with valuable information when given insight into the factors that control employee absence behaviour. Having this insight will enable the managers to promote a healthy working environment, thus lowering employee absence rates to a minimum.\n",
"role": "assistant"
}
] |
An analytical review of the factors stimulating smaller firms to export: Implications for policy-makers
|
[
"Exports",
"Small enterprises",
"Export markets"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Exporting is the most popular, quickest, and easiest way for many small companies to become international. This is because, as opposed to other modes of foreign market entry, it requires less commitment of organizational resources, offers greater flexibility of managerial actions, and involves fewer business risks[1]. Leading export and international business textbooks (Albaum et al., 2004; Czinkota and Ronkainen, 2006; Hollensen, 2004) cite numerous benefits that exporting can offer to both national governments and business organisations. From the macro-national perspective, it increases domestic employment levels, provides a source for foreign exchange, assists in the development of new technologies, creates backward and forward linkages in the economy, and leads to higher standards of living. At the micro-business level, it improves the firm's financial position, helps to develop competitive advantages, enriches managerial skills, makes better utilisation of production capacity, and facilitates company growth.Despite these benefits, a large number of small enterprises refrain from exporting, as well as from other international business activities, and prefer to concentrate only on the domestic market. Evidence provided by the US Department of Commerce (2000), for example, has shown that only a third of domestic producers is engaged in exporting, while two-thirds of the country's manufactured exports come mainly from approximately 1,000 large multinational corporations. The roots of this problem can be better understood by analysing the reasons that stimulate the initial adoption and subsequent development of export activity. In fact, export stimulation has been an issue of major concern for many international marketing researchers since the early 1970s, when the first empirical studies were conducted on the subject. Since, then, we have seen dozens of studies, conducted in various parts of the world, aiming to explain this phenomenon and identify the factors that can induce small manufacturers to export.Although insightful and useful, the knowledge gained from these studies suffers from five major flaws:1. it is non-programmatic, with some studies duplicating the work of others;2. it is widely fragmented, thus making it difficult to provide a unified body of understanding of the subject;3. it is characterised by different, inconsistent, and, in some cases, conflicting empirical results, attributed mainly to the adoption of different methodological approaches;4. it provides only a partial examination of stimulating factors, thus neglecting a number of other factors with a possible crucial motivating role in export initiation and expansion; and5. it does not offer a detailed analysis of the specific nature and importance of each stimulus on the decision to export (Leonidou, 1995a).These problems, coupled with the slowdown of this research stream in the last few years, reflect the need to consolidate the literature on the subject.As a response to this need, the aim of this paper is to critically analyse and creatively synthesise the extant knowledge on what stimulates smaller firms to export and present this in a form that would be useful to both public and business policy-makers. It is based on a review of 32 empirical studies conducted during the period 1974-2005[2] (see Appendix). The remainder of this paper is organised as follows: first, the nature of existing knowledge on export stimulation is discussed; then, the investigation method used in carrying out the review of the literature is explained; the following section offers a comprehensive analysis of the characteristics, content, and impact of each stimulus on the firm's decision to export; finally, important conclusions are drawn from the study findings and policy-making implications presented.\nExtant knowledge on export stimulation: Export stimuli, also called motives, incentives, or attention evokers, refer to all those factors triggering the decision of the firm to initiate and develop export activities. However, these factors can constitute a real driving force in exporting only to the extent with which they are brought to the attention of the key decision-maker(s) within the organisation (Wiedersheim-Paul et al., 1978). The firm's exposure to these stimuli can be seen as an organisational learning process, whereby key information is provided that determines decisions to expand internationally (Tan et al., 2007). In addition, the effective activation of an export stimulus will be subject to the facilitating or inhibiting role of various background variables associated with the environment (e.g. economic conditions, market size, infrastructural facilities), the organisation (e.g. corporate objectives/strategies, resource availability, nature of products), and the manager (e.g. management style, managerial competences/qualities, management demographics). Thus, the mere emergence or existence of an export stimulus is a necessary, but not a sufficient condition for the firm to decide to export (Dichtl et al., 1984)[3].The firm's engagement in exporting can be the result of a wide array of stimuli, which can be classified into internal and external. Internal stimuli are associated with influences endogenous to the firm (e.g. accumulation of unsold inventory, desire for extra corporate growth, possession of a unique product), while external stimuli are derived from the environment within which the firm operates or intends to operate (e.g. unsolicited orders from abroad, favourable exchange rates, encouragement by government agencies) (Simpson and Kujawa, 1974). Alternatively, export stimuli may be categorised as proactive, denoting a firm's interest in exploiting a unique organisational competence or market opportunity (e.g. possession of a financial advantage, special management interest, unutilised production capacity), or reactive, indicating engagement in export activities as a response to internal or external pressures (e.g. intense domestic competition, saturation of domestic market, stabilising sales of a seasonal product) (Johnston and Czinkota, 1982).Firms motivated by internal factors have been described as more rational- and objective-oriented in their export behaviour, compared to those stimulated by external factors. Moreover, a proactively stimulated firm is more aggressive and strategic-oriented in exporting, as opposed to those stimulated by reactive factors that adopt a rather passive and opportunistic approach towards entering foreign markets. Thus, the decision to initiate and/or expand exports can be influenced by a variety of stimuli which can be proactive-internal, proactive-external, reactive-internal, or reactive-external (Albaum et al., 2004). The impact of each stimulus on export behaviour is measured in terms of its frequency, intensity, or importance as these are perceived by managers. Usually, the thrust of the stimulating impact comes from one or a few stimuli, which will often determine whether the firm's future export actions will be systematic or sporadic.Stimuli can play an active role at any stage of the export development process, from the pre-engagement and initial stages to the more advanced and committed stages (Leonidou and Katsikeas, 1996; Morgan, 1997). However, both their nature and impact tend to differ from one stage to another, signifying their dynamic nature and adjustment according to changes in the firm's internal organisation and external environment (Crick and Chaudry, 1997). For instance, during the early export stages, the firm is more likely to be influenced by reactive and external factors (e.g. receipt of a fortuitous order from abroad), as opposed to later stages where the stimulation is more likely to be proactive and internal (e.g. desire to achieve extra organisational growth from exporting). Research has also shown that there are significant differences between exporters and non-exporters with regard to the way the various stimuli are perceived, with the former usually adopting a more favourable attitude compared to the latter (Simpson and Kujawa, 1974)[4].\nA classificatory analysis of export stimuli: The literature review revealed 40 different stimuli with a possible impact on small firms' export behaviour[5], which, for analytical purposes, were divided into internal and external (with an additional indication of whether each is proactive or reactive) (Figure 1). Internal stimuli were further divided into those related to the key functional areas of the organisation, namely, human resources, financial, research and development, production, and marketing, while external stimuli were categorised into those referring to the market (domestic and foreign), the government (home and foreign), the intermediaries, the competition, the customers, and miscellaneous. Because of variations in the results amongst the various studies, the aggregate impact of each stimulus on export initiation/development was estimated[6] (Table I). In the following, each export stimulus identified in the review is analysed in detail.Human resource-related stimuli\nConcluding remarks: The previous analysis has demonstrated that the small manufacturer can be engaged in exporting as a result of the emergence of many motivating factors in the internal and external environment, with some being of a proactive and others of a reactive nature. Each stimulus may be perceived as having a different frequency, intensity, or importance, which may further vary in accordance with time, spatial, and industry contexts. For instance, the stimulating role of competitive pressures is stronger nowadays than some decades ago, the urge to export as a result of patriotic duty may be more prevalent in Japan than in the USA, and the possession of proprietary technical knowledge appears to be more applicable to producers of industrial rather than fast-moving consumer goods. All these differences clearly indicate the dynamic, complex, and situation-specific nature of export stimuli.Irrespective of contextual factors, our review revealed that certain motives systematically play a key role in encouraging small firms to export, such as the desire to achieve extra sales, profits, and growth, exploit a unique/patented product, reduce home market dependence, and respond to unsolicited orders from abroad. Other important stimuli include: special managerial interest in exports, avoidance of the threat of a saturated domestic market, utilisation of better idle production capacity, possession of a financial advantage, need to achieve economies of scale, need to exploit proprietary technical knowledge, and identification of better opportunities abroad. Some of these motives (e.g. fortuitous orders) are characterised by an opportunistic approach to exporting, while others (e.g. sales/profit potential) exemplify a more deliberate export approach. While in the former case it is quite possible for the firm to cease or fail in its export activities on encountering the initial obstacles, in the latter case, the solid foundations that are set help to guarantee a successful exporting pattern.The review has also shown that the small firm may equally be stimulated by numerous other factors (external, internal, proactive, or reactive) that may have a lesser influence on the decision to export. Such stimuli include, for example, encouragement by external agents, smoothing product seasonality, accumulation of unsold inventory, favourable foreign exchange rates, government export incentives, relaxation of foreign restrictions, and so forth. Despite their lower motivating impact, these factors should not be underestimated for three major reasons:1. they may, under certain conditions, have a high impact on firm behaviour;2. they may complement or augment the role of the other stimuli that have a high impact; and3. they may have a serious supporting effect on the continuation of exporting once the strength of the primary export stimulus is reduced.Hence, export stimulation should be seen as the synergistic outcome of a bundle of factors which changes over time and under different conditions.\nPolicy-making implications: Several policy-making implications can be extracted from the previous review of factors stimulating exports. With regard to public policy-makers, the relatively moderate aggregate impact of government-related export stimuli implies that there is a need to:* improve existing export promotion programmes, by designing them according to the specific needs of small manufacturers and making them more attractive;* increase coordination among the various government departments and agencies in order to enhance hands-on assistance provided to would-be exporters on export procedures, export market research, and strategic marketing matters;* make various other government institutions more aggressive in turning small manufacturers into regular exporters, by providing financial, technical, or other assistance; and* facilitate firms to gradually move from exporting to other more advanced forms of internationalization, such as forming joint ventures with foreign partners.It is also important to encourage small firms to seek proactive and internal, as opposed to reactive and external, motives in their initial and subsequent export activity. This can be achieved by offering proper counselling advice and educational programmes focusing on the importance of exports in sustaining company growth and profitability, the crucial role that adequate preparation can play in beginning exporting, and the need to strategically expand into overseas markets (or even withdraw when necessary from some of these markets).Company policy-makers should also take a number of actions to enhance their firm's export potential. First, they should realise that the type and strength of the initial export stimulation will largely determine the future export behaviour of their company. Hence, they should seek to begin exporting after a thorough understanding of the underlying motives, as well as the implications that these may have on future export actions. Their approach to export markets should be based on a preliminary investigation of foreign markets and careful planning of the activities that need to be undertaken to serve these markets. The stimulation to export should be primarily based on proactive factors (e.g. fulfilment of sales, profit, and growth goals, capitalisation on competitive advantages and strengths, and systematic identification of foreign market opportunities), rather than being a reactive response to various internal or external pressures (e.g., idle production capacity, a downturn in domestic economy, and saturated domestic market). The fact that fortuitous orders from abroad is one of the most common reasons for many firms to begin exporting should be viewed with great caution by managers, since it may lead their firms into problematic export avenues. Hence, each customer order should be examined carefully to ensure that it conforms to the firm's goals for profits, sales, or growth, and its overall strategic export plans.\n",
"role": "user"
},
{
"content": "Purpose: - Although exporting can offer many benefits to smaller manufacturers, a large number of these firms refrain from export operations as a result of insufficient stimulation. This paper seeks to critically analyse and creatively synthesise the reasons that may stimulate a smaller firm to export, based on a review of 32 empirical studies conducted in various parts of the world during the period 1974-2005.\nDesign/methodology/approach: - A total of 40 export stimuli were systematically identified from the extant empirical literature, which, for analytical purposes, were divided into internal and external, as well as proactive and reactive. Within each study, stimuli were ranked in terms of their importance, frequency, or intensity, and the aggregate impact of each stimulus in all studies under review was evaluated.\nFindings: - The review revealed that export stimulation stems from a variety of factors, and may vary according to time, spatial, and industry contexts. Irrespective of contextual factors, there are certain motives that systematically play a key role in encouraging smaller firms to export, such as the desire to achieve extra sales, profits, and growth, utilise better idle production capacity, exploit a unique/patented product, avoid the threats of a saturated domestic market, reduce home market dependence, and respond to unsolicited orders from abroad. Some of these motives may lead to an opportunistic approach to exporting, while others denote deliberate export adoption. It was also shown that, although there are numerous other stimulating factors, with a lower impact on exporting, these should not be underestimated because their role may increase under certain conditions or become complementary to export stimuli with a stronger impact.\nResearch limitations/implications: - The findings of the study have serious implications for both public and company policy makers. Policy makers may use this insightful analysis of export stimulation as a guide to developing proper export promotion programmes and sound export marketing strategies.\nOriginality/value: - The paper offers a comprehensive review and synthesis of all factors with a possible stimulating effect on exporting; evaluates the aggregate effect of each factor, as collectively derived from all empirical studies conducted on the subject; and provides an in-depth analysis of the nature and the stimulating mechanism of each factor.\n",
"role": "assistant"
}
] |
Help or harm? The effects of ICTs usage on work-life balance
|
[
"ICTs usage",
"Family-to-work conflict",
"Work-family balance satisfaction",
"Family-work segmentation preference",
"Person-environment fit",
"Boundary theory"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Ubiquitous information and communication technologies (ICTs) such as the internet and smartphone have blurred the lines of demarcation between work and nonwork (Colbert et al., 2016). Employees transit between life roles more frequently and effortlessly facilitated by ICTs. For example, they can reply to friends' instant messages in the workplace once they come through on their smartphones. The steady integration of professional and personal lives is a novel workplace reality for nearly all the organizations and employees. Exploring ICTs' role on the dynamic interplay of work and family domains is a rising trend among scholars (e.g. Derks et al., 2016; Piszczek, 2017).\nTheoretical framework and hypotheses development: Nonwork-related ICTs usage, FWC and WFBS\nMethods: Sample and procedure\nResults: Convergent and discriminant validity\nDiscussion: This work aims to investigate the significant, yet not well-understood, influence of ICTs usage, exploring the implications for work-family outcomes of nonwork-related ICTs usage while at work. We investigate not only the mediating role of FWC between ICTs usage and WFBS and the moderation effect of FWSP, but also exploring these effects simultaneously. There is evidence that the influence of ICTs usage on work-family outcomes is not so straightforward as previously thought. In accordance with that evidence, our results underpin the meaning of a contingent view by investigating how one individual attribute (i.e. segmentation preference) affects the relationships between ICTs usage and FWC, and ultimately WFBS. Specifically, this study demonstrates that as nonwork-related ICTs usage during working time elevates, employees will experience augmented FWC. This augment in FWC exerts a negative effect on employees' WFBS. Additionally, the results indicate that nonwork-related ICTs usage at work predicts less FWC among those who are high in integration preference, because it is aligned with their desired boundary management styles. The results also demonstrate that the moderating role of FWSP on the relationship between nonwork-related ICTs usage during office hours and WFBS is partially mediated by FWC.\nTheoretical implications: This research contributes to the scholarly literature from several aspects. Foremost, we enrich the understanding of the effects of ICTs usage on dynamic work-family interaction in a Chinese context. Rather than focusing on work-related ICTs usage during off-job hours as pervious research does, we extend the scope to examine ICTs usage for personal reasons at work. In doing so, this work fills a formerly acknowledged gap in the research field regarding how nonwork-related ICTs usage at work can affect work-life balance. This study finds that boundary-crossing ICTs usage at work also has a positive effect on FWC. These findings are aligned with accumulated empirical research indicating that attending to work and family domains simultaneously results in boundary blurring and role conflict ultimately (e.g. Ashforth et al., 2000).\nPractical implications: Given ICTs' prevalent embeddedness in the workplace, it is impossible for organizations and employees themselves to restrict ICTs access from personal purposes completely. How to minimize FWC caused by ICTs usage and strengthen WFBS is clearly key for organizations. This work points to the significance of organizations being aware of the advantages of employees' boundary management styles. This information needs to be leveraged so as to embrace a diversity of work-family styles without stigma (Kossek, 2016) and cultivate workplace environment accommodating different styles. This is in line with Ashforth et al.' (2000) caution that employees differ in their integration/segmentation preferences and therefore organizations should not forcefully or blindly blur family and work.\nLimitations and future research: Despite the contributions, we acknowledge several potential limitations. To begin, data are collected at one-time point, thus limiting the causal interpretations. Absent longitudinal data, this study relies on theories to guide the temporal ordering of the variables. Nevertheless, we acknowledge that longitudinal data are needed to substantiate these claims more accurately. Second, reliance on self-report measures may introduce common method variance. Despite we separated the antecedent and outcome variables and guaranteed anonymity (Podsakoff et al., 2003), the bias cannot be eliminated with complete certainty. Third, by employing a non-random population sampling methodology as the recruitment strategy, some control was relinquished, allowing sampling bias and restricting generalizability potentially (Butts et al., 2015). Of particular concern is the sample composition that represents employees with relatively more wealth and work experience and better education, limiting the generalizability of our findings to other groups, for instance, less educated blue-collar workers (Liu and Li, 2018). Although the sample contains various occupations and organizations, future research can verify the hypothesized framework in more controlled settings.\n",
"role": "user"
},
{
"content": "Purpose: Leveraging boundary theory and person-environment fit theory, the purpose of this paper is to propose a model that illustrates the process by which nonwork-related information and communication technologies (ICTs) usage during office hours affects family-to-work conflict (FWC) and work-family balance satisfaction (WFBS), and the moderating effect of family-work segmentation preference (FWSP) in that process.\nDesign/methodology/approach: Analyses using PROCESS were conducted to analyze the proposed relationships on survey-collected data.\nFindings: Augmented ICTs usage led to greater perceptions of FWC, though the strength of this association was contingent on FWSP in that integration preference buffered ICTs usage's detrimental effects. Moreover, ICTs usage was shown to exert an effect on WFBS through FWC. The results indicated a mediated moderation pattern such that the moderating role of ICTs usage and FWSP on WFBS was mediated by FWC.\nOriginality/value: By focusing on ICTs usage for family reasons at work, the study filled a formerly acknowledged gap in the research field regarding how nonwork-related ICTs usage at work affected work-life balance.\n",
"role": "assistant"
}
] |
rESSuME: Employability Skills Social Media SurvEy
|
[
"Survey",
"Recruitment",
"Employability skills",
"Employer perception",
"Personal attributes",
"Social media profile"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Over the past decades, employability as a concept has shifted from a dichotomy: employable vs unemployable to a dynamic multifaced conceptualisation (Gazier, 1999). It \"should be understood as being derived from, and affected by, individual characteristics and circumstances and broader, external (social, institutional and economic) factors that influence a person's ability to get a job\" (McQuaid and Lindsay, 2005, p. 206).\n2. Interactive employability: organization - individual match: Interactive employability is the \"dynamic interaction of individual attributes, personal circumstances, labor market conditions and other context factors\" (McQuaid and Lindsay, 2005; p. 207). Within this setting, three dimensions interact with each other: individual factors, such as employability skills, attributes and demographics; personal circumstances, for instance, social capital and work culture; and external factors, for example, recruitment and labour market considerations (McQuaid and Lindsay, 2005). In terms of perspectives from which to examine employability, three are proposed: societal, which regards employment rates and economics; company/organizational, which examines the individuals and whether supply meets demand; and individual, which signals someone's opportunity to gain and keep a job (Thijssen et al., 2008). Finally, depending on the actors, the objectives of employability vary. For instance, governments strive to achieve full employment; employers aim to find the best match between the company's needs and available skills; and individuals focus on improving their career perspectives (Guilbert et al., 2016). Thus, employability interactions between employers/organizations and individuals revolve around identifying and assessing individuals' employability skills to achieve the best available match.\n3. Designing and developing rESSuME: The design and development of rESSuME involved the following phases:\n4. Results: rESSuME was delivered online to 708 directors, managers and others involved in recruitment processes in the UK and the USA. 539 completed surveys were returned. Out of all replies received, 124 were disqualified because the respondents did not meet the qualifying criteria of using FB to screen candidates. In total, 415 completed qualifying surveys were collected. The statistical analysis of the responses was conducted using IBM SPSS Statistics 24(r).\n5. Conclusion: Not hiring the right person can result in undesirable situations and negative consequences, such as low productivity, training costs, hostile work environment and others, for companies and employees alike (Waring and Buchanan, 2010). To avoid this, recruiters and employers should follow systematic screening procedures to effectively ascertain whether candidates meet the companies' needs (Macky, 2008). Knowing the personal employability skills of candidates is crucial to predict future performance (Carrigan, 2007). Although this has been extensively studied in traditional selection procedures, understanding regarding the use of SMNs in these processes is still limited (Hoek et al., 2016). While there is evidence that recruiters and employers are screening candidates SMPs (CareerBuilder, 2016) predominantly to gain information that would be difficult to access through more traditional means such as CVs (Society for Human Resource Management, 2016), there is concerned regarding the potential mismatch between the objective of SMN's content and the organizations' use of this information. Although SMNs predominantly hold personal information, which bears little relevance to job-related matters, reports indicate employers may be using this information for recruitment and job-related assessment. A further concern is the lack of guidance, training and systematic approaches to screen candidates' SMPs according to consistent and valid criteria (El Ouirdi et al., 2016, Van Iddekinge et al., 2016). This may lead employers to make bias and uninformed decisions based on available but irrelevant information.\n",
"role": "user"
},
{
"content": "Purpose: This paper presents rESSuME: Employability Skills Social Media SurvEy, which is a tool developed to understand if and how employers screen candidates' social media (SM) to identify personal employability attributes. In doing so, the purpose of this paper is to shed light into the potential mismatch between the personal purpose of SM and recruiters' job-related use of this data.\nDesign/methodology/approach: rESSuME maps personal employability attributes to elements of Facebook (FB). It was delivered to 708 employers in the UK and the USA. The 415 completed surveys were statically analysed.\nFindings: More than 75 per cent of those surveyed use FB to screen most candidates. Loyalty and reliability are the personal attributes employers most search for. They look for personal attributes examining posts, comments and photos. Country and gender differences are also reported. While in the USA, they focus on determining whether candidates have good appearance, in the UK they are more interested in gauging if candidates are reliable. Females are more concerned with establishing whether candidates display common sense than their male counterpart.\nOriginality/value: This work is the first to articulate a rationale and systematic approach to screen candidates' SMPs to: identify personal employability attributes and systematically map personal attributes to features of FB. Thus, it contributes a novel, systematic and structured tool to do so: rESSuME. It is the largest study with recruiters to date and the first to provide empirical evidence on how candidates' SMPs are screened: what personal employability attributes do recruiters looked for in SMPs; and what sections and features of FB do recruiters looked at to identify the candidates' personal employability attributes.\n",
"role": "assistant"
}
] |
CSR leaders road-map
|
[
"Corporate social responsibility",
"Leadership"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Growing challenges posed by diminishing scarce resources, increased awareness of uneven global developments, and consumption driven by particularist market forces creating a need for sustainable development all call for new forms of environmentally aware leadership. Such leadership takes organisations forward in a socially conscious manner, making decisions based not only on tangible, financial issues and short term gains, but also on intangible or non-financial issues involving sustainable growth. We term this \"corporate socially responsible (CSR) leadership\". Being a CSR leader is more than wanting to help the planet. CSR leadership is a difficult path to pursue, full of contradictions and dilemmas, where the overall destination depends on first determining a number of linking staging posts. For example, investing in social programmes requires the buy in of varying stakeholders and, in particular, shareholders who may hold different personal and organisational priorities. As Chester Barnard (1938, p. 262) observed, \"the conduct of every man is in part governed by several private moral codes\". However, \"if there exists several or many private codes governing the conduct of an individual, specific acts or concrete situations are likely to involve conflicts between codes\" (Barnard, 1938, p. 263). Thus, an individual may feel responsible for some moral codes, but not so with respect to others, particularly if they are juxtaposed to each other, as sometimes are contractual obligations (e.g. to shareholders) and moral determinations (e.g., to preservation of environment). As Chester Barnard (1938, p. 262) points out, whatever morality exists in the individual becomes effective only in conduct.\nLeading for sustainable social responsibility: On the way to gaining an understanding of CSR leadership is the need to analyse its components which, in turn, surfaces a complexity of paradoxes. The ability to run an efficient and profitable business can be consistent with a CSR agenda, but occasions will occur when the corporate financial agenda is not always consistent with stated CSR goals. In those instances, a level-headed look at integration is key to CSR goals being realised, rather than just talked about, through truly socially responsible leadership.The \"endurance of organisations depends on the quality of leaders; and that quality derives from the breadth of the morality upon which it rests\" (Barnard, 1938, p. 282). Thus, CSR leadership requires the adoption of wisdom, considered as the \"capability to identify pertinent pathways forward when direction is obscure(d)\" (Kakabadse and Kakabadse, 1999, p. 321). The direction to take may be obscured because of the different views at top team level concerning market dynamics, or because colleagues, for idiosyncratic reasons, operate from different assumption or moral standards (Kakabadse and Kakabadse, 1999). Additionally, scholars (Burns, 1978; Conger and Kanungo, 1998; Zaleznik, 1989, Kotter, 1990), have highlighted the different perspectives adopted by individuals who hold senior leadership (i.e. strategic) as against middle management (i.e. operational) roles. Such differences are displayed in terms of goal orientation, career expectations and relations with employees and people external to the organisation. Kakabadse and Kakabadse (1999) argue that the pulling together of these different forces is critical to organisational sustainability and nowhere more so than in the case of CSR.\nPurpose and method of study: This paper reports on extensive international research into CSR leadership. The study has been carried out to ensure the widest qualitative data collection possible. Case based analysis of over 65 organisations, consisting of both for profit and not-for-profit, located in the USA, UK, Continental Europe, Africa and Australia, gave a full range of issues for consideration.The research sample was determined by opportunity for access, bearing in mind the purposeful and strategic nature of the study (Gilchrist, 1992, p. 57). In the subset of 65 organisations where the authors had access to board directors, CSR concerns were explored from the perspective of those directors. The interviews focused on the relevance of CSR to boardroom discourse. In other organisations, interviews were conducted across management levels, and CSR adoption was examined from the early stages of policy determination at board level to application at the operational levels of customer/community interface. This ensured that data collection was not only world-wide but also corporate-wide in nature. All levels of staff were considered in the total process and well over 300 people were interviewed for this study. Interviews, data feedback, and participant observation were the particular methodologies adopted.With such a range of companies, such a range of countries, and such a range of people the intensive qualitative approach had every chance of capturing the type of information needed to isolate the skills a truly effective CSR leader would need.Once the interviews were complete, a thematic analysis of the rich, qualitative data was conducted, gleaned from 300 interviews. Presented in this paper is the pattern of themes that recurred across participants' narratives (Riessman, 1993; Pentland, 1999). Whilst all human knowledge creation is a process of interpretation, \"interpreting is a continuous monitoring process that takes place before, during and after a research performance\", highlighting the fact that qualitative data analysis is itself a process of construction.All interviews were tape-recorded and transcribed. The narratives scripts were subsequently coded adopting a grounded theory approach, as emerging themes and frameworks were amended in the light of the subsequent transcriptions (Glaser and Strauss, 1965). The transcripts were read independently by all inquirers, identifying recurring narratives and staying close to original meanings (Miles and Huberman, 1984; Guba and Lincoln, 1994). In this way, the creation of abstract categories from each individual leader's experience was minimised.\nStudy results: Regardless of the country of residence or size of the company, research findings revealed three stages of CSR implementation, common to all, where particular and different CSR leadership skills would be needed. The decision-making stage is the initial stage, at which the individual leader or leaders come to terms with the need for personal and organizational movement towards CSR. The adoption stage follows the decision-making stage, and is the stage where most of the leadership capabilities the respondents mentioned would be required. This is the stage at which real forward movement begins and, if effective, starts to spread into and impact the CSR leader's organization. Finally, a commitment stage was evident in terms of a real signing up to CSR, rather than merely going through its process. This stage is reached, for individuals and for organizations, when it is clear that the CSR goals will be consistently pursued despite obstacles such as shareholder objections, conflicting priorities, difficulties of measurement, and the like.From the interview data, we also identified ten specific CSR leadership skills and capabilities that appear to have relevance in one of the three stages discussed above. Each skill appeared to have its own place on the road to achieving CSR, and together they represent a coherent whole that provides the means for driving CSR through organisations. These are all represented in Table I.While these capabilities are important for any CSR leader to have or to develop, one might envision that some of the skills might also be dispersed across or within a highly effective CSR team or department, and that in an organisation with a highly developed CSR application, all leaders would have each of the requisite skills, and would be able to use them appropriately and with the best timing.Stage One: CSR decision\nImplications and conclusions: The model presented here represents the journey to effective CSR implementation taken by the organisations we studied. It is based on two equally important components:1. the stages through which the CSR leader progresses, both as an individual and with his or her organisation (if the leadership skills are developed and the leadership actions are effective), from initial uncertainty and ambiguity in understanding CSR, to moving forward to a new and concrete reality framed by deliberate leadership action; and2. the skills and capabilities needed, or required to develop, at each of these stages, for successful movement forward of organisational CSR.As shown in Figure 1, the skills are interconnected, in the sense that all need to be employed at many points over the course of time, and in that they need to be often used together to face inevitable challenges confronting the implementation of CSR. The focus is on highlighting both the leadership challenges in CSR and the management tasks and objectives - the interlocking of the need for individual leader capabilities and the movement or culture change needed in most organisations for full implementation of CSR.As emphasised, truly potent CSR leadership requires the individual responsible to be able to energise all ten skills so that these come together, allowing organisational leaders to act in a coherent manner that makes sense to the outside world and that drives the organisation forward. A lack of or misuse of any one of the ten \"travel points\" of CSR leadership has potency to damage and undermine the effectiveness of others along the way. For example, should the leader have not fully thought through his/her case, then the persuasion of management is undermined and, in so doing, the business case for CSR and the language adopted are weakened. Under such circumstances, the tensions and paradoxes in the organisational system multiply, damaging CSR measurement and, in turn, consistency of application and hence follow through. Such experience undermines the will to act, display and live the commitment to CSR at both the level of the leader and the organisation - a particularly unwelcome phenomenon - for stage III acts as the window to the rest of the world. Stage 3 emphasises what is publicly evident, what the organisation will and will not do, can and cannot do. Should certain skills be neglected, that becomes evident, first to the organisation, and second, to the external world. Should the decision to adopt CSR be more driven by concern for external image improvement, such positioning is evident to the managers, whose motivation to seriously pursue CSR through disciplined measurement, consistency and application and follow through, is likely to be undermined.Although, in some ways, the synergy arising from the ten unified aspects of leading CSR may be perceived by observers as being borne of innately charismatic leadership qualities, such serendipitous construction can be deliberately crafted by individuals who have a passionate desire to be CSR leaders. CSR leadership can be developed. Yet learning to master each aspect and then perform the ultimate alchemy is a challenge for leaders, as well as for active leadership development.So these stages and skills, as well as their interconnectivity and impact, require deliberate consideration by the leader and by his or her organisation. Unaware of how one skill affects the other damages CSR adoption. Awareness of the intermingling of leadership and organisational concerns prepares all for ensuring that CSR initiatives can be made to work.Yet CSR awareness and good intent is not sufficient. Discontent arises from good intent going sour! All too often, the hard emotional labour of managing paradoxes, conflicts and clashes means organisational demands compete with CSR goals and initial good intent never reaches the bridge of follow through which allows for crossover to action. This, in turn, can create both a fear of CSR as well as an urge to understand CSR issues in more depth. By deepening their CSR awareness and enhancing their abilities to engage in reflexivity, leaders also nurture a level of leadership wisdom that accommodates the handling of paradox. Through contemplation and the re-focusing of one's own energy, as well as that of the organization, the CSR leader can help bring about the unification of all ten aspects into a harmonious, coherent and consistent leadership push. Connectivity that provides essence or energy creates the internal desire to incite action. Once this is accomplished, CSR goals can be achieved through a talented and creative synthesis which reconciles the tensions of organisational life and ensures CSR to become the gene of the organisation's DNA.\n",
"role": "user"
},
{
"content": "Purpose: - The purpose of this paper is to outline how CSR can be effectively implemented and driven through the organisation. The emphasis of the paper is not on CSR definition but on the skills and capabilities needed by individuals and organizations to fully implement CSR application.\nDesign/methodology/approach: - As a qualitative study, interview, data feedback, and participant observation were the particular methodologies adopted.\nFindings: - Three stages of CSR implementation and, within those stages, ten leadership skills and capabilities are identified. The nature of their inter-relationship and how that impacts on application, is discussed and explored. The ten skills and capabilities form a portfolio for individual leaders to consider and indeed develop in their management of CSR. These capabilities are called forth in three logical stages from those required for early decision making to those required for full enactment of CSR, forming a clear model. This model provides a road-map for leaders to increase their consciousness and their effectiveness in the implementation of true rather than token CSR.\nResearch limitations/implications: - A limitation is the qualitative case-based method. The learning arising from the study can be pursued and further tested through quantitative survey methods in order to provide for balanced, comparative analysis.\nPractical implications: - A road-map to effective CSR application for leaders of organisations is offered. This road-map can be used to guide current leaders and as a guide to developing future leaders.\nOriginality/value: - Originality is high as no such model of CSR application exists. The value of the paper is to offer a research-based practical guide to CSR implementation.\n",
"role": "assistant"
}
] |
Process improvement capability: a study of the development of practice(s)
|
[
"Organizational structure",
"Critical success factors",
"Higher education",
"Process improvement",
"Action Research"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: UK higher education institutions (HEIs) were already competing in a market environment before significant changes to the UK funding mechanisms occurred in 2011 (Gibbons, 2005; Carpentier, 2006; Hemsley-Brown and Oplatka, 2010). (Bryde and Leighton, 2009) suggest these commercial dynamics had already propelled HEIs to focus on effectiveness and efficiency, also identified by others (Coates and Mahat, 2014; Milliken and Colohan, 2004). The Universities UK Group's report about efficiency and effectiveness made recommendations that universities need to \"continue to prioritise streamlining internal and operational processes\" (UK Efficiency and Modernisation Task Group, 2010, p. 35), and Bendermacher et al. (2017, p. 40) articulate a similar argument that universities \"need to invest in organisational change processes in order to be able to gain position in an increasingly competitive market\". This paper responds to the call by Bendermacher et al, for empirical research into how quality practices can be developed which support continuous improvement of organisational processes. This is aligned to the contention by Ahmad et al. (2007) that business process management (BPM) re-engineering initiatives in HEIs need to employ an organisational development element (Ahmad et al., 2007) as well as Klun and Trkman's (2018) call for BPM researchers need to develop more in-depth theory that understands processes more as organisational routines. This response is the development of \"mode 2\" knowledge (Viet et al., 2006) (mode 1 being scientific, mode 2 being social, applied, cross-disciplinary, etc.) through an intervention research (IR) project, which is a particular mode of action research (Baard, 2010) to consider process improvement (PI) routines. We pose the following research question and sub-questions:\nStep 2: literature review: The literature review is segregated into two elements aligned to the two research sub-questions:\nStep 3: creating the construct: The PI Unit had already run two pilot projects, and experimented with some other interventions, including what are sometimes termed \"Rapid Improvement Events\". These type of interventions have been applied in public sector, including HEIs (Radnor, 2010; Radnor and Bucci, 2011). The stated priority for the unit was to develop capacity and competence in the wider team in running PI projects. However, the pilot projects had been led just by the Intervention Researcher and the Unit manager. Hence, it was agreed that the intervention to be constructed and tested was a PI project led by a member of the Unit. This, therefore, would be the arena for studying the development of PI routines. The project was based in a student management office located in the University, and included members from the office and related departments. The project had been scoped and agreed following an Improvement Event run by the Unit almost a year previously. The applied construct for the scope of the research design was an improvement meeting, preceded by one of the Unit's \"normal\" operating/planning meetings.\nStep 4: implementing and testing the construct: As discussed earlier, this investigation into routine development adopts the practice perspective. One methodology that is seen as particularly useful for investigating practice is that of MDA (Nicolini, 2012). Scollon and De Saint-Georges (2001) consider MDA as \"a form of action research, intimately bound to the specifics of situation studied and issue researched\" (p.15). MDA is also part of the ethnographic tradition (Scollon, 2001a) which itself has been used for empirical studies into routines (Howard-Grenville, 2005). It has its roots in the activity theory and can be tracked back to the sociologist Vygostky (Norris and Jones, 2005). The activity theory and hence MDA assumes that all social actions are mediated through tools, external artefacts or internal (to the individual) tools. MDA shares some tenets with critical discourse analysis; however, in contrast, MDA has a distinct focus on action, and sees discourse as just one among many potential mediational means. Of particular relevance is the inclusion, within these means, of material objects, such as artefacts, which as previously noted, are highly pertinent to the development of routines.\nStep 4 continued: testing the construct: research design and data collection: The pivotal activity in MDA is located in observation of the actual mediated actions at the site(s) of engagement. The site of engagement in this case was the second meeting of the PI project team. Scollon (2001a) is clear that MDA assumes no priori assumptions about which mediational means or social actions will be important. Hence, a form of pre-event data collection was used to establish the potential mediational means, the trajectory of relevant practices, and any other significant sites of engagement for these practices and participants (de Saint-Georges, 2005). Triangulation, and hence a degree of validity, is attempted in MDA by capturing different types of data, including generalisations of members, objective observations, individual members experience and observers interaction with members. Data were collected from observations (and additional interaction with participants), and a set of post-event interviews, followed by the sharing of the initial analysis with a group of the participants, referred to within MDA as a focus group. Figure 2 shows a summary of overall methodological design which has been constructed using the methodological steps outlined in Scollon (2001a), for the designated site of engagement.\nStep 5: findings and analysis: The next step within CRA is to analyse the findings with a view to establishing the theoretical contribution. MDA provides a range of heuristic questions (Scollon, 2001a), and researchers Jones et al. (2017) utilise these to develop an analytical pathway; this was used to provide structure for this analysis.\nStep 5 continued: theoretical development: Two key aspects of theorizing are the notion of abduction and retroduction (O'Mahoney and Vincent 2014). Abduction is theoretical re-description of events often using theory gleaned from the literature review in order to obtain the most plausible explanation of events (O'Mahoney and Vincent, 2014; Bygstad et al, 2016). Retroduction is the \"what-if\" creative analysis to identify potentially hidden mechanisms. Mingers(2006) and Ketokivi and Mantere (2010) both argue that abduction and retroduction are often done in \"one movement\" by researchers as they move from qualitative data to the best explanation of the data. O'Mahoney and Vincent (2014, p. 16) succinctly describe this process as \"adding theory to data\", and this is seen as compatible with both IR in general and CRA in particular (Jonsson and Lukka, 2005). Lynham (2002), in her discussion on theory building in applied disciplines, outlines that the conceptual framework developed must be broken down into elements such as propositions built from operationalising the theory from the data. Jonsson and Lukka (2005) confirm that the target of the exploratory IR such as CRA is to develop new theoretical propositions. In a synthesis of these two approaches, proposition sets, which are essentially candidate mechanisms, will be produced as a result of the abduction and retroduction process.\nStep 6: conclusions and examination of applicability: The latter steps of CRA involve excising the theoretical contribution and examining the scope of potential applicability and generalisability. This is mainly focussed on the applicability of the identified mechanisms, with some consideration of the similarity of the context in which the outcomes occur (Ackroyd and Karlsson, 2014). In order to provide substantiation for the relevance of these PI routines beyond the higher education sector, these were tallied with the wider PI and BPM literature. As to be expected, there were numerous references to these practices, so three indicative, illustrative studies are identified for each routine, shown in Table AII This provides a degree of substantiation that the routines investigated have some potential generalisability outside the HEI sector, to any organisations attempting to develop PI capability.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to report on a research project, using intervention research (IR), which aims to identify how a higher education institution could develop process improvement (PI) capability.\nDesign/methodology/approach: The paper adopts a practice perspectives of routines, and classifies and catalogues the potential routines that could form PI capability. The development of these routines are investigated using the constructive research approach, a form of IR), in the action research mode. Within this approach, the methodology of mediated discourse analysis was employed to trace the empirical trajectory of the routine development, in a student management office within the context of an improvement project by the institutions PI unit.\nFindings: Of relative significance is the implication that there is a small group of initialising PI practices which are accessible to practitioners, in contrast to a large set of critical success factors. Second, these PI practices transcend particular methodologies, meaning their development can be incorporated into customised, contextualised methodologies, by individual organisations.\nPractical implications: The set of PI practices identified are able to be enacted by practitioners and are not dependent on macro-management factors. Second they are relatively simple to understand and are not associated with any particular improvement fad or fashion.\nOriginality/value: The study contributes to the appreciation of PI in higher education as a capability, and outlines the potential array of routines that could constitute that capability. It provides a theoretical view on how key PI routines are developed in an organisational field, and a more nuanced and richer view of \"process mapping\" and its effect on other PI practices.\n",
"role": "assistant"
}
] |
Statistical comparison of final weight scores in quality function deployment (QFD) studies
|
[
"Quality function deployment",
"House of quality",
"Parametric bootstrapping",
"Relationship matrix"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1 Introduction: Quality function deployment (QFD) is a methodology used to translate the \"voice of the customer\" (VOC) into engineering and technical specifications to be followed in the design of products or services. Akao (1990) has reported that when appropriately applied, QFD has been effective in substantially reducing product development lead times. The main goal in implementing QFD is to improve the quality of the product or service based on customer-defined requirements and expectations. Although QFD is a popular and widely used technique, as Enriquez et al. (2004 cited in Garver, 2012) point out, on-going research still seeks to examine the assumptions and methods used within QFD with a view to continuously improving the methodology and there is a need to be able to accurately determine importance scores for the customer because with inaccurate data \"the entire House of quality (HOQ) is built upon a weak foundation\" (Garver, 2012).\n2 QFD and its factors: In developing a HOQ, the customer, competitor and engineering data that populate the matrices and vectors are of an inherently qualitative nature, and are operationalised into numerical values through rating scales that transform linguistic criteria into numeric data. A wide variety of practice is observable in the application of these linguistic-numeric scales. In the rating of customer priority, competitor position, etc. there is not only potential variability in determining which value on a given scale most closely aligns with the perceived \"reality\"; but there is also wide variation in the scales that are applied by practitioners. In the following section we explicate the commonly used linguistic-numeric scales and outline their use in QFD.\n3 Methodology: testing of FW differences using a parametric bootstrap method: 3.1 Monte Carlo testing\n4 Results: 4.1 Determining the significance of differences between extreme FW ratings\n5 Conclusions: In relation to the first aim of the research, in this paper we have demonstrated that not all of the differences between the FWs of HOW attributes may be significant. Indeed, for one of our literature-derived examples (Tan et al., 1998) we have demonstrated that in the context of common cause variation, even the most extreme HOW FWs are not significantly different from each other. This finding implies that the engineering attributes necessary to maximise customer satisfaction may, in the course of a QFD analysis, be prioritised inappropriately, and action may be taken in respect of one HOW requirement in preference to another, where there is in fact no statistical difference between their ratings. A practical implication of this is that organisations may engage in costly or time consuming activity resulting from the prioritisation of an engineering attribute, where an attribute requiring less effort or cost may be an equal priority.\n",
"role": "user"
},
{
"content": "Purpose: Quality function deployment (QFD) is a methodology to translate the \"voice of the customer\" into engineering/technical specifications (HOWs) to be followed in designing of products or services. For the method to be effective, QFD practitioners need to be able to accurately differentiate between the final weights (FWs) that have been assigned to HOWs in the house of quality matrix. The paper aims to introduce a statistical testing procedure to determine whether the FWs of HOWs are significantly different and investigate the robustness of different rating scales used in QFD practice in contributing to these differences.\nDesign/methodology/approach: Using a range of published QFD examples, the paper uses a parametric bootstrap testing procedure to test the significance of the differences between the FWs by generating simulated random samples based on a theoretical probability model. The paper then determines the significance or otherwise of the differences between: the two most extreme FWs and all pairs of FWs. Finally, the paper checks the robustness of different attribute rating scales (linear vs non-linear) in the context of these testing procedures.\nFindings: The paper demonstrates that not all of the differences that exist between the FWs of HOW attributes are in fact significant. In the absence of such a procedure, there is no reliable analytical basis for QFD practitioners to determine whether FWs are significantly different, and they may wrongly prioritise one engineering attribute over another.\nOriginality/value: This is the first article to test the significance of the differences between FWs of HOWs and to determine the robustness of different strength of scales used in relationship matrix.\n",
"role": "assistant"
}
] |
Farm types and adoption of proven innovative practices in smallholder bean farming in Angonia district of Mozambique
|
[
"Typology",
"Mozambique",
"Smallholder farmers",
"Multivariate analysis",
"Bean production",
"Management and marketing"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Improved crop production requires the development and dissemination of an array of proven technologies tailored to farm-specific conditions, rather than blanket recommendations across diverse farms (Chikowo et al., 2014; Zingore et al., 2011). Previous research for Sub-Sahara Africa (SSA) suggests a high degree of variability among smallholder farmers in terms of both socio-economic and bio-physical conditions (Chamberlin, 2007; Chikowo et al., 2014; Tittonell et al., 2010). Chikowo et al. (2014) even reiterated that, household wealth, priorities, and crop production objectives, determine to a greater extent the availability and allocation of resources to different activities. It therefore implies that adoption and rates of adoption of technologies will vary between farms of dissimilar resource endowment and production orientation, leading to variation in crop productivity at the farm level. More so, it implies that, technological interventions expected to yield better results in fighting poor productivity in smallholder farming systems must be designed to target these diverse and heterogeneous farms and farming systems (Tittonell et al., 2010). Failure to recognise this heterogeneity is often linked to projects yielding lower than expected impacts in SSA (Chikowo et al., 2014; Giller et al., 2011).\n2. Agriculture and socioeconomic development in Mozambique: The agricultural sector plays a crucial role in the socioeconomic development of Mozambique. The majority of the population living in rural areas heavily depend on agriculture for their livelihood (Lopes, 2010; Lukanu et al., 2004). The sector accounts for an estimated 28 per cent of the country's gross domestic product (Tomo, 2009) and employs over 75 per cent of the population (Lopes, 2010; Tomo, 2009). Subsistence agriculture is the dominant employer of the workforce though agricultural productivity is weak (Cunguara et al., 2011). The agricultural sector in Mozambique is dualistic, comprising of commercial and smallholder sub-sectors. The smallholder sub-sector cultivates a large share of the land and produces most of the food crops (Mango et al., 2015).\n3. Bean production, technology development and dissemination in Mozambique: Common bean is an important legume in human diets globally; providing protein, iron, zinc, fibre and complex carbohydrates (Gepts et al., 2008) and food for over 400 million people in Africa. In Eastern and Southern Africa, beans are recognised as the second most important source of human dietary protein and third most important source of calories (Buruchara, 2007). As food pulses, beans are a cheap but important source of high-quality proteins, and are consumed throughout the year since they are easy to store. They are, thus, one of the best means of mitigating food nutrition problems experienced in countries like Mozambique. Furthermore, common bean is also a very important source of income for smallholder farmers. The production of common bean in Mozambique is on the increase, with an increase of about 55 per cent evident between 2002 and 2012 (Trabalho de Inquerito Agricola (TIA), 2012). In Figure 1, we show the trends in bean production by province from 2002 to 2012.\n4. Materials and methods: 4.1 Description of study area\n5. Results and discussions: 5.1 PCA results\n6. Conclusions and policy implications: We rely on a multivariate analysis approach that combines PCA, and CA to identify five distinctive farm households within Angonia district of Mozambique with respect to the adoption of proven innovative practices (technologies and methods) in smallholder bean farming, using socio-economic factors. We analysed data on 30 variables from 322 smallholder bean farmers. PCA identified ten factors that accounted for nearly 63.45 per cent of the variance in the original 30 variables. The identified ten factors were used in CA to characterise smallholder bean farming households. Using the ward technique and an expert knowledge of the study area, CA led to the creation of five farm types. The first farm type is characterised by female landowners with small farm sizes who have low rates of adoption of improved bean seed, inoculant use, topdressing fertilisers and certified seed use. The second type represents educated farmers with access to credit. The group has the highest rates in adoption of inoculants, basal and top-dressing fertilisers, use of improved and certified bean seed and participates in the bean market as a collective unit. The third type represents rich male landowners with large farm sizes but low education levels. This group has high rates in adoption of improved and certified seed, high average weeding frequency and use of topdressing fertilisers. However, collective market participation, use of inoculants and basal fertilisers is poor in this group. The fourth type represents the young and inexperienced poor male farmers who do not use improve bean seed, inoculants in production or collective market participation. The group only has relatively higher rates in use of fertilisers and weeding frequency. The fifth type is characterised by experienced female farmers with high labour endowment who have low rates of adoption of improved bean seed and do not use inoculants. Moreover, the group has an above average rate of fertiliser use and high average rate in practising collective market participation.\n",
"role": "user"
},
{
"content": "Purpose: Improving the adoption rates of proven innovative practices in bean farming and their impacts on livelihoods requires persistent promotion of practices, complemented by rigorous socioeconomic analysis that recognises the diversity of smallholder farmers. The purpose of this paper is to typify farm households in Angonia district of Mozambique, based on their socioeconomic characteristics prompting the adoption of proven innovative practices in bean production, management, and marketing.\nDesign/methodology/approach: The authors use a multivariate statistical analysis approach that combines principal component analysis, and cluster analysis to clearly identify five distinctive farm household types with respect to the adoption of proven innovative practices in smallholder bean farming using socio-economic factors.\nFindings: The study findings show that various socioeconomic factors define clusters and can be associated with the adoption and use of innovative practices in smallholder bean farming. The five farm types identified are: female landowners with small farm sizes (29.52 per cent); educated farmers with access to credit (6.63 per cent); relatively rich male land owners with large farm sizes and low education (8.73 per cent); youthful, inexperienced and poor male farmers (6.33 per cent); and experienced female farmers with high labour endowments (8.43 per cent). The respective farm types seemed to have different patterns in the adoption of proven innovative practices in bean farming.\nOriginality/value: The authors recommend that policy makers promote strategies meant to raise adoption of innovative practices in bean production, management and marketing in Mozambique that takes into account household diversity. The farm types identified by this study can be a good starting point for guiding such future efforts.\n",
"role": "assistant"
}
] |
A risk-control framework for e-marketplace participation: the findings of seven cases
|
[
"Electronic commerce",
"Risk analysis"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: The increasing trend in the use of internet-based e-marketplace applications has created tremendous opportunities for businesses to manage effective supply chain management. IDC predicts IT and e-marketplace spending will reach $496.7 billion in the USA and $1.3 trillion globally by the year 2009. White and Daniel (2003) describe e-marketplaces as web-based systems that enable automated transactions, trading, or collaboration between business partners. According to Bakos (1998), an electronic marketplace is an \"inter-organizational system that allows participating buyers and sellers to exchange information about processes, products, and services.\" Table I presents the definitions of e-marketplaces from several different sources. We adapt Bakos (1998) definition of e-marketplace as \"an Internet IT platform which facilitates many buyers and suppliers to exchange business transactions, goods and services.\" Online procurement, which once was the domain of a proprietary firm's extranet, is being transformed by the growth of e-marketplaces, trading hubs and online auctions (Grishankar, 2000).E-marketplaces have grown rapidly in recent years and have received considerable attention in e-business research (Bakos and Brynjolfsson, 1999; Palmer et al., 2000). They attract a larger pool of buyers and suppliers who are transacting electronically to negotiate for the best price in order to increase liquidity and have a variety of price mechanisms leading to online auctions (Dai and Kauffman, 2001). E-marketplaces have had a large amount of hype and are continuing to grow, even though the past records of e-marketplaces were full of uncertainties, vulnerabilities and risks. The exposure to risks increased when disparate services were provided by the various trading partners (Choudhury et al., 1998; Sims and Standing, 2002). The new services were threatened by internal factors such as; the lack of standards, lack of regulations, and the lack of secure systems. For example, although Covisint provided a unique framework called \"industry open systems\" which was designed for collaborative e-marketplace in the automotive industry, however, it suffered from anti-trust risks due to the lack of uniform standards which were not included in their initial regulatory approvals. Ultimately, Covisint lost sight of their electronic platform mission and began undertaking commercial alternatives which led to competition in various input products and service markets. Furthermore, external factors such as volatile online political sanctions, natural hazards, legal issues, environmental issues, and other political instabilities threatened the firms using their framework. This study examines e-marketplace security. We identified four types of risks namely; economic, technological, implementation and relational risks. The seven e-marketplace firms were chosen arbitrarily from several different industries. The four types of control implemented by the firms lead to the development of a risk-control framework based on the findings of this study. The study contributes to practitioners as it increases the importance and awareness of e-marketplace security. Table I presents the definitions of e-marketplaces from different sources.Previous research identified several distinct types of e-marketplaces including buyer-driven, seller-driven, independent vertical and horizontal marketplaces that serve numerous purposes as presented in Table II (Bailey and Bakos, 1997; Brooks and Dik, 2001; Christiaanse and Markus, 2003; Cuny and Richardson, 2001; Kaplan and Sawhney, 2000; Lenz et al., 2002).\nRisks in e-marketplaces: E-marketplaces provide the basic infrastructure to allow suppliers and buyers to interact in an online environment. Despite its growth risks associated with e-marketplaces were predominantly caused by weak operating procedures that evolved during the development process, deficiencies in e-marketplace protocols, and a myriad of technology related problems (Kaihla, 2002; Vaidyanathan and Devaraj, 2003). Further, a host of e-marketplace risks arose from accidental or erroneous processing of business transactions. For example, services were threatened by internal factors due to lack of standards, regulations and/or an efficient support system. This was particularly evident in a collaborative environment where both external and internal business partners are logging into the system (Christiaanse and Markus, 2002). Reshaur and Turner (2000) suggest that risks can be viewed as hazards, uncertain outcomes, or missed opportunities.Economic risks\nResearch method: A multiple case study research approach was applied in this study. A random sample of e-marketplace firms were selected in order to provide a cross section of diverse industries including; the automotive, aerospace and defense, chemicals, construction, energy, agriculture and plastics industries. Since, e-marketplace firms are internet-enabled, the sample was selected by initially conducting a web site analysis in order to identify firms that would be appropriate for this study. We then called the IT managers of the firms we selected and explained the purpose of this study. The goal of this study was to examine risks that firms encountered along with the controls they utilized in e-marketplace B2B e-commerce. Those firms who agreed to participate in this study were sent a two-part e-mail questionnaire. The first part asked for them to rate the most important types of risks they experienced according to the four categories we listed. The second part concerned the controls measures the firms implemented in those four categories. The primary source of data collection was gathered via telephone interviews. Each telephone interview lasted for 30 minutes and we spoke to three employees from each firm at various times. On average, we spent 2 hours per telephone interview, followed by e-mail clarifications of the data collected. Using a Likert scale where low was 0-3, medium 4-6, and high 7-10, we paid particular attention to their responses that rated the impact of their risks and controls. In order to maintain the privacy and confidentiality of the participants, we agreed not to use their names and their quotes.Secondary background information of the firms was obtained from their web sites. This information included the details of their products, business partners, type of industry and other background information. To gain an even better understanding of the firms in this study, we reviewed academic journals, press releases and practitioner journals that provided the business strategies and philosophies.Table III presents the background characteristics of the seven firms that participated in this study. Additional description of the background information of the firms is included in the Appendix.\nFindings of the risks in e-marketplaces: In this section, we discuss the findings of risks from the four categories namely; economic, technological, implementation and relational risks that the firms experienced. Then we discuss the findings of the controls from the four categories that the firms implemented.Economic risks\nFindings of the controls of e-marketplaces: Controls refer to measures taken by the e-marketplace firms to manage, mitigate and eliminate their risks. Control is a part of a facility, including any system, procedure, process, or device that is intended to eliminate hazards or prevent hazardous incidents from occurring. Controls can also reduce or mitigate the severity of consequences should any incidents occur. The e-mail questionnaire that was sent to our participants had a section pertaining to the control measures that the firms enforced based on the risks that they had experienced in each of the risk categories. The firms undertook a risk analysis assessment in order to identify the controls that will reduce and mitigate the risks they had experienced. Some of the firms such as Exostar, Farms.com, ChemConnect and Covisint applied an operational resilience program to identify the areas where risks occurs and applied the right controls. The category of control implemented depended on the severity and impact of the risks. We named the four categories of controls economic, technological, implementation and relational to match the four types of risks.Economic controls\nThe impact of risks and controls: Table IV presents the impact of the risks and the controls through a quantitative analysis and applying a Likert scale. The details of the risks and controls are discussed in the findings section. This section draws upon the impact of the risks and controls the firms experienced. The impact levels were measured using a ten-point Likert scale with low (0-3), medium (4-6), and high (7-10). Our findings suggest that although e-marketplace security was a concern for most of the firms, they were able to control it with effective IT solutions, standard operating procedures and regulations.As participants were describing the types of risks they experienced during the telephone interviews, we also requested them to indicate the impact it had on the risks and controls. The quantitative analysis suggests that the impact of the risks were mostly medium to high as most of these firms were late adopters and did experience integration and compatibility issues that led to implementation, relational and economic risks from unwanted IT investments.The impact of the controls implemented by the firms were mostly rated medium to high. One reason for this is that there were some firms such as Covisint and ChemConnect who have implemented e-marketplace earlier and had a lot of data whereas, others were late adopters and their IT managers did not experience significant risks to date in order for them to enforce controls pertaining to the risks. The products and services the firms had were unique to a specific industry type and a group of suppliers. These firms included; Exostar in the aerospace industry, Farms.com and Construction.com.\nThe risk-control framework for e-marketplaces: Technology is evolving at such a rapid pace that those firms who want to survive must conduct business through e-marketplace. The down side to having such a fast paced evolution is that security systems cannot always keep pace with the global systems that must be developed to conduct business world-wide. Based on the results of our research concerning the risks experienced and controls developed by the seven firms we contacted, we formulated a risk control framework for e-marketplace participation which is presented in Table V. The framework lists the key types of risks encountered along with the corresponding controls utilized by the seven firms in our research.To elaborate on one example in the automotive industry, suppliers experienced security risks when their competitors were able to view proprietary designs transmitted online, standard risks when they purchased expensive and complicated software that had to be replaced with new systems, implementation risks when the software, online exchanges, and other digital initiatives failed to perform as expected, early-adapter risks when a well-intended supplier invests in a system and becomes the guinea pig or beta test for the rest of the industry, opportunity risks when they spend millions or billions of dollars on software and technology risk when the money would have been better spent on core necessities such as factory upgrades or wage increases.Covisint was among the first e-marketplaces to be launched. They had no other models to follow and were restricted by the limitations of technology. ChemConnect rated their relational risks to be high due to the Enron case and they had fears of credit risks. These risks coupled with the economic downturn put ChemConnect under pressure to ensure that trading on its exchange would be fair. Table V, presents the risk-control framework for e-marketplace participation.\nConclusions: This study examined the role of e-marketplace security in seven firms across diverse industries. We identified four types of risks and controls namely; technological, economical, implementation, and relational risks and controls. Most of the economic risks experienced by the firms came from the competitors, thereby forcing e-marketplace firms to focus on quality and specialized services. In order for business partners to sustain long-term relationships, trust becomes important and firms that trust their business partners are capable of making effective decisions. Lack of an effective security framework can be disastrous and costly, not only leading to a loss of business revenue but also adversely impacting the integrity of the firm's e-commerce operations and reputation. The findings contributed to risks factors in e-marketplaces based on four categories. Likewise, the findings also contributed to control measures that e-marketplace firms can implement leading to the development of a risk control framework that highlights the types of risks experienced and controls implemented by these firms thereby increasing the awareness of practitioners. Further, the findings contribute to theory as it provides a generic framework that included vital factors that impact e-marketplace risks and controls. Most of the firms in our research realized the economic benefits from e-marketplace participation and have planned projects to develop internet security guidance documentation. Their participation will also provide guidance for addressing internet security as part of an overall site vulnerability assessment, help in developing process control security practices and standards, and insight toward implementing a voluntary benchmarking standard. Future research will be able to apply the risk control framework in a different context.\n",
"role": "user"
},
{
"content": "Purpose: - E-marketplaces have recently become a popular internet-based e-commerce application. Despite extensive research on this topic, there has been limited work in the realm of e-marketplace security. This paper aims to discuss the role of e-marketplace risks and controls in seven e-marketplace firms.\nDesign/methodology/approach: - The authors have identified four types of risks and controls of e-marketplaces namely; technological, economical, implementation, and relational. Seven e-marketplace firms from a cross-section of industries were chosen.\nFindings: - The research suggests that the type and extent of risks, as well as the type of industry and the products the firm sells, have a direct correlation with the method of e-commerce application implemented for e-marketplace.\nOriginality/value: - The findings contribute to a risk control framework in e-marketplace participation which will increase the awareness of security in e-marketplace participation.\n",
"role": "assistant"
}
] |
Servitization of the manufacturing firm: Exploring the operations practices and technologies that deliver advanced services
|
[
"Case study",
"Servitization",
"Product-service systems",
"Operations",
"Service delivery systems"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Manufacturing and service industries are often assumed incommensurable. Whether discussing national economies, business classifications, education, training or employment, they tend to be thought of as separate. Yet manufacturers themselves can base their competitive strategies on services, and the process through which this is achieved is commonly known as servitization (though servitization and servicisation are often used interchangeably) (Wise and Baumgartner, 1999; Oliva and Kallenberg, 2003; Slack, 2005; Schmenner, 2009). This strategy can strengthen customer relationships, create new and resilient revenue streams, and set high barriers for competitors (Baines et al., 2009b; 2011a).\nBackground and context: Concept of servitization and advanced services\nMethodology: Overall approach to the research\nKey findings: Facilities and their location\nDiscussion: In this section we set out to assimilate our findings and consider the wider aspects of this study.\nConclusions and future work: Contribution\n",
"role": "user"
},
{
"content": "Purpose: This paper aims to explore practices and technologies successfully servitised manufacturers employ in the delivery of advanced services.\nDesign/methodology/approach: A case study methodology is applied across four manufacturing organisations successful in servitization. Through interviews with personnel across host manufacturers, their partners, and key customers, extensive data are collected about service delivery systems. Analyses identify convergence in their practices and technologies.\nFindings: Six distinct technologies and practices are revealed: facilities and their location, micro-vertical integration and supplier relationships, information and communication technologies (ICTs), performance measurement and value demonstration, people deployment and their skills, and business processes and customer relationships. These are then combined in an integrative framework that illustrates how operations are configured to successfully deliver advanced services.\nResearch limitations/implications: The analyses are reductive and rationalising. Future studies could identify other technologies and practices. Case study as a method is inherently limited in the extent to which findings can be generalised.\nPractical implications: Awareness and interest in servitization is growing, yet adoption of a servitization strategy requires particular organisational capabilities on the part of the manufacturer. This study identifies technologies and practices that underpin these capabilities.\nOriginality/value: This paper contributes to the understanding of the servitization process and, in particular, the implications to broader operations of the firm.\n",
"role": "assistant"
}
] |
Deception and lies in business strategy
|
[
"Strategy",
"Homophily",
"Fake news",
"Case vignettes",
"Deceit",
"Typologies of deception"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: Fake news has become a hot topic of public debate all over the world. Fake news is pervasive, spanning traditional media, such as broadcast television, to social media. What is fake news by definition? There does not seem to be a universally accepted definition. In the field of communications study, a typical definition of fake news would be that it is a genre of programming where either the program's central focus or a very specific and well-defined portion is devoted to political satire. An example would be satirical skits on television which portrays politicians as self-centred, dumb, lecherous and dishonest. Fake news can be perceived as unrealistic while traditional (hard) news is perceived as realistic and this in turn may cause viewers to weigh the content of fake news differently from that of hard news.\nFake news and deception: Outside the domain of academic literature, finding a definition is not easy. For example, a recent news article (The Wall Street Journal, 2016) states that Facebook Inc., Twitter Inc. and Alphabet Inc.'s Google are under pressure to expunge fake news, making them reluctant judges of what is not true or misleading. This perhaps gives a layman's definition (i.e. not true or misleading) of fake news. Something that is \"not true\" would be akin to lying, while \"misleading\" would be akin to deception. What is the distinction between lying and deception? Lying, according to Merriam-Webster (Merriam-Webster, 2016), is deliberately stating what is false with the intention of deceiving. In comparison, deception is when someone intentionally misleads the receiver into a belief which the deceiver believes to be false. The deceiver can deceive someone without asserting something. For example, a deceptive advertisement can be one which sheds favourable light on a product using selected data that in turn make the audience come to false conclusions about the product (Bivins, 2008). The advertisers are not lying because they did not offer any conclusions about that product. This is the difference between lying and deceiving.\nDeception in business: The primary purpose of this paper is to conceptualise deception in business to offer a framework that provides typologies that allow categorisation of the common forms of deception from known case studies. Through the use of case studies, we can discern the motivation for the use of a strategy of deception by businesses. Towards this purpose, this paper adopts Hopper and Bell's (1984) six typologies that act as determinants of categories of deception. The typologies are then applied to real case study vignettes. Hence, each vignette is then categorised to fit the typologies seen as the most appropriate.\nConclusion: The paper explores the use of deceit by businesses using Hopper and Bell's (1984) typologies of deception with several case vignettes. Each case vignette provides a narrative which explains the events and an analysis of the deceitful actions. Through the analysis it is possible to then fit each case vignette into one or more of the typologies of deception. The vignettes offer examples of the following typologies: crime, masks, lies and unlies. This paper makes a contribution by showing how Hopper and Bell's typology applies to real business situations. The paper raises the question as to whether businesses use deceit as a strategy.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to explore the use of deceit as a business strategy. There is ample evidence in the mainstream media of deceitful strategies in business, yet there is a lack of discovery of deceit as a strategic tool in the mainstream literature.\nDesign/methodology/approach: This is a conceptual paper that first explains deceit and interprets the use of deceit as strategic tool in business using case vignettes as evidence. The paper puts forth a convincing case that there is enough evidence to substantiate our proposition that deceit is indeed part of the repertoire of tools utilised by some businesses.\nFindings: The value of this paper is that it highlights why deceit is used strategically to achieve profit motives of businesses.\nOriginality/value: This paper attempts to fill a gap that exists in the extant literature and would especially benefit management practitioners and business academics in appreciating the use of deceit as a business strategy.\n",
"role": "assistant"
}
] |
Lords of Huh?
|
[
"Corporate strategy",
"Experience"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n__NO_TITLE__: I just finished quite an interesting book about the history of management consulting and the \"business strategy revolution\" with which it coincided. It is Walter Kiechel III's The Lords of Strategy: The Secret Intellectual History of the Corporate World (Harvard Business Press). Kiechel is a former Editorial Director of Harvard Business Publishing, and a former managing editor at Fortune Magazine.\n",
"role": "user"
},
{
"content": "Purpose: - Review and discussion of the book Lords of Strategy by Walter Kiechel III.\nDesign/methodology/approach: - Opinion column.\nFindings: - \"Strategy\" has not been seen as something in which businesses must engage. The \"Strategy revolution\" Kiechel outlines arose in the 1960s with the new breed of consulting firm, had its heyday then and in the 1970s, and declined thereafter into something like incoherence and a focus on tactical, financial, and cost-cutting approaches.\nOriginality/value: - Reviews an important book and extracts some lessons about business strategy.\n",
"role": "assistant"
}
] |
Corporate governance and payout policy: evidence from India
|
[
"Family firms",
"Corporate governance",
"Dividend policy"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Dividend is the distribution of residual profit among the shareholders, the suppliers of finance. Elucidating dividend policy has been one of the most backbreaker challenges faced by financial economists. Despite numerous studies, it is still difficult to understand the factors which influence dividend policy, and how these factors interact with each other. More than 40 years ago, Black (1976) wrote, \"the harder we look at the dividend picture, the more it seems like a puzzle, with pieces that just don't fit together\" (p. 5). The situation is pretty much the same today. In a recent study of dividend policy, \"much more empirical and theoretical research on the subject of dividends is required before a consensus can be reached\" (Allen and Michaely, 1995, p. 833). Dividend acts as a governance mechanism to control manager-shareholder agency problem, as it decrease the available cash, which could potentially be expropriated otherwise (Easterbrook, 1984; Faccio et al., 2001). According to agency theory, managers have an opportunity to use corporate resources in ways that benefit their own interests and not the funds suppliant (Jensen, 1986). Paying higher dividend to shareholders subsides the magnitude of agency conflict (Ben-Nasr, 2015; Firth et al., 2016). Rozeff (1982) introduced agency cost as likely explanation of dividend policy and found significant relationship between dividend policy and agency costs. The question arises then how a firm's director can be persuaded to pay higher dividends when their personal inclination is to retain cash.\n2. Literature review and hypothesis development: 2.1 Board composition and dividend policy\n3. Sample description and methodology: 3.1 Sample\n4. Conclusion and implications: The study focuses on the Indian capital market with its unique setting due to implementations of corporate governance mechanism and dominance of family governed businesses. Using panel data techniques on a sample of 1,536 firms, the study reveals three major findings that are robust to alternative model specifications, Tobit and Probit model estimations.\n",
"role": "user"
},
{
"content": "Purpose: The purpose of this paper is to study the impact of ownership structure and corporate governance on dividend policy in emerging markets, like India. The study also analyses the moderation effects of board independence between ownership and dividend payout.\nDesign/methodology/approach: The data set of 1,546 Indian firms over the period of 2006-2017 has been used in this study. Tobit and logistic regression methods has been used. The data used in this study are collected from the Centre for Monitoring Indian Economy (CMIE) Prowess database. The sample firms are listed on Bombay Stock Exchange (BSE) and National Stock Exchange (NSE).\nFindings: First, the study finds a significant positive influence of corporate governance on the decision to pay dividend and is an important determinant of the payout decision. Second, the study finds a significant negative relationship of family ownership with dividend payout decisions which indicates that family firms pay lower dividend. Finally, the result from the interaction effect of board independence with family ownership has significant positive influence on dividend policy.\nOriginality/value: This is one of the first attempt to show that there is an interaction between independent board and ownership structure. It shows that more independent and non-executive directors in the board of family controlled firms are likely to pay more dividends.\n",
"role": "assistant"
}
] |
Educational data mining: a systematic review of research and emerging trends
|
[
"Educational data mining",
"Learning analytics",
"Systematic review",
"Prediction of performance",
"Decision support",
"Artificial intelligence"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\n1. Introduction: Technology innovation is transforming the world at an astonishing speed, including the field of education. The development of learning technologies has enabled teachers to track student learning in fully online, blended, as well as brick and mortar learning environments by recording student behaviors via technological means (Hung et al., 2017; Dimic et al., 2019). By analyzing the vast amount of educational data that are generated and collected in the process of teaching and learning, decision-makers such as instructors, students and administrators can obtain a holistic view of learning progress and trigger corresponding evidence-based or data-based personalized interventions or recommendations (Hung et al., 2017). These areas of research are called educational data mining (EDM) and learning analytics (LA), aiming to extract useful knowledge for supporting teaching and learning (Siemens and Baker, 2012). EDM and LA have been the focus of intensive research in the past decade.\n2. Literature review: Based on a search of the following five digital databases, including Web of Science, Proceedings of the Educational Data Mining conference, the Journal of Educational Data Mining, ACM Digital Library and IEEE Xplore, 1,219 articles and proceedings related to \"educational data mining\" were searched. These 1,219 articles and proceedings were published from 2007 to the first quarter of 2019. In addition, 33 review articles were identified within the vast number of publications, and a basic overview of those review articles were reported in the following subsection.\n3. Research method: 3.1 Data search procedure\n4. Results: 4.1 Research trends\n5. Discussion and open issues: 5.1 Research trends\n6. Conclusion: The systematic analysis results of existing EDM publications enable us to draw several conclusions. A first insight would be that EDM publications are mainly contributed by actual analysis studies. In addition to basic statistics and regression, machine learning and even deep learning algorithms have been widely used in EDM. However, EDM research mainly focuses on higher education rather than K-12, and collecting small data sets for EDM research is predominant, especially in K-12 field. Four major EDM research topics, including prediction of performance, decision support for teachers and learners, detection of behaviors and learner modeling and comparison or optimization of algorithms, were identified. Some interesting findings and open issues related to these research topics have been reported and discussed deeply. It is also concluded that the state-of-art of EDM research is far from the ideal of AI, and the automatic support part for teaching and learning in EDM may need improvement in the future work.\n",
"role": "user"
},
{
"content": "Purpose: Educational data mining (EDM) and learning analytics, which are highly related subjects but have different definitions and focuses, have enabled instructors to obtain a holistic view of student progress and trigger corresponding decision-making. Furthermore, the automation part of EDM is closer to the concept of artificial intelligence. Due to the wide applications of artificial intelligence in assorted fields, the authors are curious about the state-of-art of related applications in Education.\nDesign/methodology/approach: This study focused on systematically reviewing 1,219 EDM studies that were searched from five digital databases based on a strict search procedure. Although 33 reviews were attempted to synthesize research literature, several research gaps were identified. A comprehensive and systematic review report is needed to show us: what research trends can be revealed and what major research topics and open issues are existed in EDM research.\nFindings: Results show that the EDM research has moved toward the early majority stage; EDM publications are mainly contributed by \"actual analysis\" category; machine learning or even deep learning algorithms have been widely adopted, but collecting actual larger data sets for EDM research is rare, especially in K-12. Four major research topics, including prediction of performance, decision support for teachers and learners, detection of behaviors and learner modeling and comparison or optimization of algorithms, have been identified. Some open issues and future research directions in EDM field are also put forward.\nResearch limitations/implications: Limitations for this search method include the likelihood of missing EDM research that was not captured through these portals.\nOriginality/value: This systematic review has not only reported the research trends of EDM but also discussed open issues to direct future research. Finally, it is concluded that the state-of-art of EDM research is far from the ideal of artificial intelligence and the automatic support part for teaching and learning in EDM may need improvement in the future work.\n",
"role": "assistant"
}
] |
Profiling European traditional food consumers
|
[
"Europe",
"Food products",
"Product image",
"Consumer psychology",
"Demographics"
] |
[
{
"content": "Summarize the following paper into structured abstract.\n\nIntroduction: Novel foods and modifications of existing food products are continuously being introduced in the European food market. Examples of product modifications are new combinations of ingredients, new production methods and/or new packaging. At such a time, when there is a strong focus on innovation, traditional food products (TFP) can be considered as \"the odd man out\". Consumers have quite broad and diverse conceptualisations as to what constitutes TFP. This conceptualisation varies between countries, cultures and even individuals. In a qualitative study, the following exploratory, consumer-based, definition of TFP was formulated:[...] a product frequently consumed or associated with specific celebrations and/or seasons, normally transmitted from one generation to another, made accurately in a specific way according to the gastronomic heritage, with little or no processing/manipulation, distinguished and known because of its sensory properties and associated with a certain local area, region or country (Guerrero et al., 2009).TFP continue to retain their market share and are even experiencing increased interest and demand (Kuznesof et al., 1997; Skuras and Dimara, 2004) from both researchers and consumers. Many research studies have discussed the underlying dimensions of the concept of traditional food (Cayot, 2007; Guerrero et al., 2009; Jordana, 2000), consumers' attitudes, perceptions and expectations towards this food category (Guerrero, 2001; Kuznesof et al., 1997; Roininen et al., 2006; Sanzo et al., 2003), its sensory characteristics and experience (Caporale and Monteleone, 2001, 2006; Cayot, 2007), the marketing opportunities from information provision related to TFP (Caporale and Monteleone, 2001; Caporale and Monteleone, 2004; Iaccarino et al., 2006) or the challenges that TFP face in contemporary food markets (Jordana, 2000). However, no studies have concentrated directly on profiling the TFP consumer. As a result, no clear-cut description of TFP consumers, in terms of socio-demographical, attitudinal and behavioural characteristics, is currently available. The existing information in the literature about the traditional food consumer (TFC) profile originates from indirect measurements in studies, which vary in terms of the country where studies were performed and in terms of the studies' goals, methods and sample characteristics.In general, traditional food consumption is more strongly common in Southern Europe than in Northern Europe. The reasons for this include a greater concentration of smaller food businesses (Lienhardt, 2004) and a wider variety of production systems and end products in Southern Europe, resulting in traditional products having a greater significance and economic weight (Jordana, 2000). The literature shows consistent relationships between the profile of TFC and the socio-demographical characteristics such as age, gender and income. Typical TFP consumers are found to be native (Kuznesof et al., 1997), older (Dagevos, 2005; Kuznesof et al., 1997; Platania and Privitera, 2006; Teagasc, 2004), female (Platania and Privitera, 2006), and living on a below average income (Kuznesof et al., 1997; Teagasc, 2004). Further research shows that rural people have more interest in TFP (Guerrero et al., 2009; Weatherell et al., 2003), as do people with a lower educational level (Teagasc, 2004).The literature also shows some associations between a TFC and values, beliefs and life-style orientations. TFC are generally not caught up in modern ways of life (Guerrero et al., 2009), where time-pressure, business and convenience orientation dominate, nor do they show much interest in other aspects of modern life such as extensive travelling, product innovations, high-tech products and new media (Kuznesof et al., 1997). Housewives are typically portrayed in the literature as typical consumers (and purchasers) of traditional foods (Guerrero et al., 2009; Platania and Privitera, 2006; Teagasc, 2004). TFC are also portrayed as liking the familiar, one expression of this preference being the consumption of familiar dishes (Dagevos, 2005; Guerrero et al., 2009; Platania and Privitera, 2006; Teagasc, 2004). This image closely relates to a preference for continuity, rather than change, in life (Dagevos, 2005; Platania and Privitera, 2006; Teagasc, 2004). In this vein, Dagevos (2005) stated that TFC have fairly conservative food habits, maintaining culinary customs across generations. Consumers of traditional foods are also considered to be people who are concerned about their health (Teagasc, 2004).However, the literature provides conflicting information about the effect of price on purchases and consumption of TFP. Dagevos (2005) has defined a TFC as a materialistic individual with a cost-conscious consumption pattern, whereas Platania and Privitera (2006) argues that price is not an important product attribute when buying TFP. Lengard et al. (2008) found that a low price was incompatible with the image of traditional food in France, Poland and Norway.TFC are also characterised by the high importance they attach to tasty and unique food products (Guerrero et al., 2009; Platania and Privitera, 2006). The origin of food has also been discussed as an important issue for TFC (Platania and Privitera, 2006). This is closely linked to what has been referred to as ethnocentrism, which involves consumers preferring domestic products to foreign ones, even when there is no obvious or straightforward reason for this (Shimp and Sharma, 1987). Given this link of TFC with ethnocentrism, we can also assume that the positive association between ethnocentrism and family life (as opposed to living alone) (Chryssochoidis et al., 2007) also helps define TFC.This study explores the profile of a European TFC invoking the theory of self-image congruity (Sirgy, 1982). This theory is based upon the reasoning that consumers do not consume products just for their material utilities, but also consume the symbolic meaning of those products as a portrayal of their self-images (Elliott, 1997). This theory implies that consumers are more likely to consume products whose images are congruent with their personal self-image. The concept of personality image plays a crucial role within this theory. A product acquires a personality image through its association with human personality. As applied in this study, the theory suggests that a consumer of TFP will report a strong congruence between their projected image of a typical TFC and his/her own characteristics.The objective of this study is to define and validate the profile of a consumer of TFP through using a pan-European consumer sample. First, a description will be given of the extent to which consumers perceive themselves to be consumers of TFP. This will lead to the distinction of two consumer segments: TFC segment and non-TFC (NTFC) segment. Second, we will discuss the images of people and/or personalities (character profiles) that respondents associate with a typical consumer of TFP (projected image). Third, the actual individual characteristics that correlate significantly with the consumption of TFP will be identified. Finally, we will check for congruity between the projected image of a typical TFC/NTFC and the actual, observed characteristics of TFC and NTFC in our sample and use this to build and define a descriptive profile of a TFC.\nMaterials and methods: Research approach and sampling\nEmpirical results: Self-reported consumption of traditional food\nDiscussion: Figure 1 shows that a majority of respondents reported themselves to be \"above average\" consumers of TFP. The respondents were randomly recruited from a consumer panel, and it is unlikely that there was any sample bias towards TFC. More likely explanations of the high mean value might be found in the positive connotation that traditional food has among consumers, and the rather vague and subjective consumer definition of TFP (Vanhonacker et al., 2008; Guerrero et al., 2009). The higher (self-reported) consumption of traditional foods in the south and the lower consumption in the north is consistent with earlier conclusions by Lienhardt (2004) and Jordana (2000).The typical projected image that European consumers have of consumers of TFP appeared to be strongly determined by notions of locality, a positive attitude towards food preparation and consumption and a traditional way of living, in which stability is preferred to change and excitement, which closely fits with the role model of a traditional family. In contrast, consumers with a typical non-traditional food consumption pattern were seen as younger people, singles, with less available time, who live busy lives, travel quite often and do not prioritise time for food shopping and for cooking.The projected image of a TFC held by people who consume a lot of TFP (TFC segment) is particularly relevant for product differentiation and marketing communication. In addition, the projected image of TFC among those who report a low consumption of traditional food (NTFC segment) might also be of interest in revealing some of the underlying reasons for their low TFP consumption. The TFC segment indicated associations with more character profiles from the list than consumers of non-traditional foods. This may suggest that the more a person eats TFP, the more he/she can imagine anyone else also eating TFP. People in both segments identified or associated themselves as being connoisseurs, health-concerned, couples with no children at home and price conscious people. People in both segments identified consumers of non-traditional foods as likely to be convenience-oriented people, frequent travellers, single and busy.Our findings are validated by the theory of self-image congruity, which combines the projected image that European consumers have about a typical consumer of traditional food (second subsection of the Section \"Empirical results\") with the actual characteristics or profile of a European TFC (third subsection of the Section \"Empirical results\"). The theory of self-image congruity implies that the character profiles that people within the TFC segment strongly associate with a typical TFC should match the actual characteristics of those within this segment. And, similarly the character profiles that people within the NTFC segment associate with NTFC should match the actual characteristics of this segment. Table VI shows the match (or not) between the perceived (image) and actual characteristics of the two segments. In terms of socio-demographics, there was strong match for gender, marital status and the presence of children in the household. While the survey did not contain any direct questions about whether TFC consumers were housewives, personal characteristics such as \"spending a lot of time on preparing food\" and \"experiencing making dinner as taking care for the family\" are clearly typical of housewives' (or house-husbands') behaviour. Finally, and perhaps surprisingly, the association of people living in the countryside with a TFP consumer was not confirmed (although not negated) at the pan-European level.Regarding life-style orientations, attitude and behaviour towards food, the self-image congruity theory is confirmed through direct and/or indirect measurements for people that are old fashioned and health concerned (associated with traditional food consumption) and for busy and convenience-oriented people (associated with non-traditional food consumption). The confirmation for old-fashioned people is based on a higher percentage of older people and the search for familiarity in food, while the concern for health was demonstrated via several direct measurements (see Table V). Although no direct measure for \"being busy\" was included in the survey, this criteria is based upon the assumption of a positive relation between higher education and being more busy. Convenience orientation is confirmed indirectly since those in the NTFC segment indicated that they sought ready-to-eat food and/or food that is easy to plan, buy, prepare and cook. Furthermore, a higher level of ethnocentrism is linked to people preferring national and regional cuisines, while enjoyment from cooking and enthusiasm for cuisine was reflected in statements such as \"trying to impress others with one's cooking skills\", \"use creativity when preparing food\" and \"consider making dinner as mainly pleasant\". In addition, the many in the TFC group read cooking magazines, confirming the association of connoisseurs with TFC.\nConclusion: This paper draws upon a large, pan-European, survey to profile the characteristics of European consumers of TFP. It generally confirms many of the elements found in recent literature with respect to the images of traditional food and traditional food consumption. In addition, the paper used the theory of self-image congruity to validate the profiles of TFC and NTFC.In summary, European TFC are typified as coming from middle-aged to older age groups, whose origins are in their country of residence. They tend to live in larger households, are health concerned, prefer familiarity and stability in their lives and in their food consumption. They enjoy and spend much time in purchasing and preparing food and perceive preparing food as a moral but enjoyable obligation towards their family. This profile of characteristics provides some opportunities for product positioning and marketing communication related to traditional foods. TFC are receptive to labels that guarantee the origin of the raw materials, to the use of organic raw materials, new processes that improve safety, packaging that preserves sensory qualities and is re-sealable, year-long availability, and more variety in each category of TFP.In contrast to TFC, people in the NTFC segment are typically younger, weight-conscious and identify more with the modern way of life, where time is in short supply and everything has to be experienced. They enjoy trying new foods, yet give less priority to shopping or preparing meals. As such they seek for value for money, time efficiency and convenience in food purchasing and preparation. They would welcome innovations in TFP, especially those that offered new flavours, shapes and textures, and changes of this nature might lead them to increase their consumption of traditional foods.These results are particularly useful for product marketing and future product development in the traditional food sector, as they provide an indication as how it can better reach and satisfy existing TFC, and possibly, adapt TFP in order to attract new segments and a new generation of consumers.\n",
"role": "user"
},
{
"content": "Purpose: - The paper seeks to provide a picture of the profile of European traditional food consumers (TFC) in terms of their socio-demographics, attitudes, life-style orientations and behavioural characteristics.\nDesign/methodology/approach: - Cross-sectional data were collected through a pan-European consumer survey (n=4,828) conducted in November and December 2007, with samples representative for gender, age and region, collected from six European countries: Spain, Italy, France, Belgium, Poland and Norway. First, information about the consumption of traditional food and the image of a typical TFC was provided through descriptive analysis. Next, the actual characteristics of TFC were assessed using partial least squares regression and these were discussed in the context of the theory of self-image congruity.\nFindings: - Traditional food consumption patterns are stronger in the south than in the north of Europe. TFC across Europe are typically middle-aged to elderly, health-conscious, ethnocentric, food connoisseurs, who are attached to familiarity in their food choices and who very much enjoy cooking.\nPractical implications: - The information about the profile, the image and the actual characteristics of TFC has practical implications for marketing and communicating about traditional foods, including their identification, differentiation, research and development and positioning in the European food market.\nOriginality/value: - Traditional food is currently undergoing a revival, with increased sales, consumer and regulatory interest. This paper provides cross-cultural results from a large pan-European consumer sample that provides highly relevant and useful information about the market potentials of the traditional food industry, which mostly consists of small-and medium-sized enterprises. The application of the self-image congruity theory to validate the results contributes to the originality of the paper.\n",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.