Spaces:
Runtime error
Runtime error
| {"docstore/metadata": {"878bdaad-e099-44e7-8cfb-d214c9e407d1": {"doc_hash": "8ad3e580478e226164372eb6d3f0038dbd2a55f4c75a17681a7f7a0dc9793305"}, "45838104-6e6e-4c29-beda-59d5941d912e": {"doc_hash": "1726317792dc3b040dabf4d551d9dff525e6b06b58e024529e0eb93106e886a6"}, "bfa53f84-e81b-4d85-80a8-28e0864774aa": {"doc_hash": "771f0d8053019949b356542730efc955873b26aa7fd13256d820ac0669daee3d"}, "9015c73c-a129-4422-bb81-2ad60b3e0458": {"doc_hash": "da74f35db0f361856268c4971ee62c81b41a2a76b5dc905de3fc44d9f85e6e74"}, "3e0c2955-89ea-4a88-ab65-323f73c022a6": {"doc_hash": "82f5548dc3e33700854bfc21389ebbcfb153f20d4c08ba5ec87a45fe2d8d66fc"}, "eb6f75f5-4119-4a47-80f1-f967c1299210": {"doc_hash": "5b3f3ae0619668d33d2d15218851766f9bf75404fa0e9bca108f75ea1287db27"}, "e8b520f8-96ef-4abe-b6f5-fcb26be295e7": {"doc_hash": "41ac1a95af4be8d599f5220fd59befdebe356d9d3950212fbd7de6d1a921cc6c"}, "041abddf-0d22-4833-a7aa-45330b5bb49e": {"doc_hash": "b589fbc40da9cbad2ac9a29578a0dd17adb08a7f41cd8f23434f4cce232c9d85"}, "eec8c5c1-025d-4d48-8c95-02960991813f": {"doc_hash": "9605d2c8f2cc0c372b29e7e113b7a25c2236950a29e295e3bad68e986ab8e169"}, "5c196df3-e04a-46bc-89a1-d3406c2f0357": {"doc_hash": "c954a2aab1e96c45105df6cb42564aa1a918c35259b2f21093ed6095e0a43d32"}, "76d2b8ea-8fc5-4e16-8413-1e632581c4d7": {"doc_hash": "36d339883c37d21f760fee15228d35e8c8b938f3ee9c6267aab5be6265b342d8"}, "6695aaa4-3f68-4863-9eff-8e37406349e8": {"doc_hash": "e133cabca104cddd8f0adf1f7a9ad54f360b39d4acb42fdcc1f6ab5d7317ca46"}, "f7c1761b-4d8d-41a2-85eb-b597ef039513": {"doc_hash": "feb29eb1299f6ccd83e275f91d4880cff76617d58c247b2c4d71c102282e4edd"}, "0780cde1-9a19-4cbd-9525-12a03b247a78": {"doc_hash": "095f2375d10187d2c1173ea12893c2419c51167771c13debe896a2ecf9e37816"}, "6012fe95-dfd4-477c-b0c5-98be7f19251a": {"doc_hash": "01eab58188ce2f2e31bbecdc4f945f59e1910edf9e2ec4f284bba450965cb5f9"}, "9aa77e4d-8cdc-4d30-98f1-053ec73fcb4a": {"doc_hash": "2f2c6bbdd753e9b91656d525c47c7ce63470678033397425231436b5d18ab5cd"}, "f7b61543-9a3f-4a9e-908a-ebf80a05d53e": {"doc_hash": "330b041a424b4099694487e34da27b691ff0ad79c73dc09daa2dac9decbcb70b"}, "c5242644-4e19-4413-9b0f-20b838b08201": {"doc_hash": "c6275dfe81b828294b7f07fe6471694c37fe5a9ec51b67079f593bc9f71db73d"}, "fb85b6f9-c436-44bb-a256-53b372536df4": {"doc_hash": "a7c9c327eaa28c1a37685a7405789b3a4fcd275c0c77b8a15637fb7d3994b06f"}, "25027da7-f677-499d-90e0-5665fa5133c5": {"doc_hash": "f53f49e06053ca0c480ae781ff6177a173e3f22217f1969f62f1bc5ba098db3d"}, "e62b0869-d6f1-4863-9be0-173f9fbfae9d": {"doc_hash": "688488b63effcb96f645323bed3441480a9d9ac0aec1b9cad761890e6a14c33c"}, "f2a49d34-de3c-4aa1-8081-4ef7b9e79a60": {"doc_hash": "eb2e1fe28f046e882d541656cb9923f4adb8044a85c857d88950e044c9e41f77"}, "bf73a3f1-3f4f-4c5d-8809-5acbede2d600": {"doc_hash": "292cdeb011353a6ad2ed315a65f8556f35cc6cbdb8d3497519257890f429292e"}, "ee3a3718-eeac-4615-808c-897a216b27a5": {"doc_hash": "e40fcbd51a88772e814351ed641ac753eff39b3b1d34e6de77941100423b5660"}}, "docstore/data": {"45838104-6e6e-4c29-beda-59d5941d912e": {"__data__": {"text": " \nA Review of ChatGPT Applications in Education, Mark eting, Software \nEngineering, and Healthcare: Benefits, Drawbacks, a nd Research Directions \n \n Mohammad Fraiwan \u0002,* and Natheer Khasawneh \u0004 \n \u0002 Department of Computer Engineering, Jordan Univer sity of Science and Technology, \nP.O. Box 3030, Irbid 22110, Jordan; mafraiwan@just. edu.jo \n \u0004 Department of Software Engineering, Jordan Univer sity of Science and Technology, \nP.O. Box 3030, Irbid 22110, Jordan; natheer@just.ed u.jo \n*Correspondence: mafraiwan@just.edu.jo \n \n \nAbstract: ChatGPT is a type of artificial intelligence langu age model that uses deep \nlearning algorithms to generate human-like response s to text-based prompts. The introduction \nof the latest ChatGPT version in November of 2022 h as caused shockwaves in the industrial and \nacademic communities for its powerful capabilities, plethora of possible applications, and the \ngreat possibility for abuse. At the time of writing this work, several other language models (e.g., \nGoogle\u2019s Bard and Meta\u2019s LLaMA) just came out in an attempt to get a foothold in the vast \npossible market. These models have the ability to r evolutionize the way we interact with \ncomputers and have potential applications in many f ields, including education, software \nengineering, healthcare, and marketing. In this pap er, we will discuss the possible applications, \ndrawbacks, and research directions using advanced l anguage Chatbots (e.g., ChatGPT) in each of \nthese fields. We first start with a brief introduct ion and the development timeline of artificial \nintelligence-based language models, then we go thro ugh possible applications of such models, \nafter that we discuss the limitations and drawbacks of the current technological state of the art, \nand finally we point out future possible research d irections. \nKeywords: Artificial Intelligence; ChatGPT; Chatbot; Machine Learning; Natural Language \nProcessing \n \n \n1 Introduction \n \nChatGPT is a type of artificial intelligence (AI) l anguage model that uses deep learning \nalgorithms to generate human-like responses to text -based prompts. The introduction of the \nlatest ChatGPT version in November of 2022 has caus ed shockwaves in the industrial and \nacademic communities for its powerful capabilities, plethora of possible applications, and the \ngreat possibility for abuse. At the time of writing this work, several other language models (e.g., \nGoogle\u2019s Bard and Meta\u2019s LLaMA) just came out in an a ttempt to get a foothold in the vast \npossible market. These models have the ability to r evolutionize the way we interact with \ncomputers and have potential applications in many f ields, including education, software \nengineering, healthcare, and marketing. \nHistorically, language models have been around for more than 20 years with some \nattempts go back to the 1960\u2019s. However, recent dev elopments in deep learning AI, the huge \ncomputational power offered by graphical processing units (GPUs), and the accessibility to large \ndatasets have enabled amazing advancements in the c apabilities and the likelihood to human \noperators. Moreover, Chatbots are now being touted as the future of search engines, because \nthey are able to formulate answers to queries rathe r than just point out the links to possible \nanswers. For example, instead of searching for prog ramming tutorials or a lucky similar answer \nto a homework assignment, ChatGPT can easily provid e the necessary code in response to such \na query with varying degrees of complexity. ChatGPT was able to pass bar exams, United States \nmedical licensing exams, and job interviews, among others [1]. \nThe first language models appeared in the 1960s. ELI ZA was developed by Joseph \nWeizenbaum as one of the first Chatbot programs [2] . It used pattern matching and pre-written \nresponses to simulate conversation with a human use", "doc_id": "45838104-6e6e-4c29-beda-59d5941d912e", "embedding": null, "doc_hash": "1726317792dc3b040dabf4d551d9dff525e6b06b58e024529e0eb93106e886a6", "extra_info": null, "node_info": {"start": 0, "end": 3896}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "3": "bfa53f84-e81b-4d85-80a8-28e0864774aa"}}, "__type__": "1"}, "bfa53f84-e81b-4d85-80a8-28e0864774aa": {"__data__": {"text": "Moreover, Chatbots are now being touted as the future of search engines, because \nthey are able to formulate answers to queries rathe r than just point out the links to possible \nanswers. For example, instead of searching for prog ramming tutorials or a lucky similar answer \nto a homework assignment, ChatGPT can easily provid e the necessary code in response to such \na query with varying degrees of complexity. ChatGPT was able to pass bar exams, United States \nmedical licensing exams, and job interviews, among others [1]. \nThe first language models appeared in the 1960s. ELI ZA was developed by Joseph \nWeizenbaum as one of the first Chatbot programs [2] . It used pattern matching and pre-written \nresponses to simulate conversation with a human use r. Fast forward to the 1990s, the artificial \nlinguistic Internet computer entity (ALICE) was deve loped by Richard Wallace. ALICE was another \nearly Chatbot program. It used a similar approach t o ELIZA, but also incorporated machine \nlearning to improve its responses over time [3]. A decade later, Cleverbot was introduced by Rollo \nCarpenter [4]. Cleverbot was a Chatbot program that used artificial neural networks to learn from \nits conversations with users. It was able to genera te more natural and varied responses than \nearlier Chatbots. More recently, in 2018, OpenAI re leased the first version of their Generative \nPre-trained Transformer (GPT) language model. It us ed unsupervised learning to train on large \namounts of text data and could generate coherent an d diverse text based on a given prompt. In \n2019, OpenAI released an improved version of their GPT model called GPT-2. It had 1.5 billion \nparameters, making it one of the largest language m odels at the time. GPT-2 was able to generate \nhigh-quality text that was difficult to distinguish from human writing. In 2020, OpenAI released \nan even more powerful version of their GPT model ca lled GPT-3. It had 175 billion parameters, \nmaking it the largest language model to date. GPT-3 was able to perform a wide range of language \ntasks, including language translation, content gene ration, and answering questions [5]. \nMoreover, the number of users of its service was re ported by media outlets to exceed 100 Million \nin two months after its launch. The latest release, ChatGPT-4 is scheduled for release in mid \nMarch, 2023. Moreover, at the time of writing this paper, Microsoft just released Visual ChatGPT, \nwhich extends the capabilities of ChatGPT by allowi ng sending/receiving images in the chat \ndialogue. \nChatGPTs are built on top of the GPT models develop ed by OpenAI, but with additional \ntraining and customization for conversational appli cations. They represent the cutting edge of AI \nlanguage technology and have the potential to revol utionize the way we interact with computers \nand each other. On February 6th, 2023, Google relea ses their own conversational AI called Bard, \nwhich was quickly followed by Meta\u2019s Large Language M odel Meta AI (LLaMA) on February 24th, \n2023. Other less known models do exist in the liter ature, including bidirectional, extreme \nmultilingual language model (XLNet) [6], GShard (a T ransformer-based deep learning \narchitecture) [7], robustly optimized BERT pretrain ing approach (RoBERTa) [8], and text-to-text \ntransfer transformer (T5) [9]. These AI language mo dels are all based on the transformer \narchitecture and have achieved impressive results i n various natural language processing tasks. \nHowever, each model has its own strengths and weakn esses, and the most appropriate model \ndepends on the specific application and the data av ailable. \nThese powerful language models represent a technolo gical disruption to the current \nacademic and industrial landscape. They may render many existing technologies (e.g., traditional \nsearch engines) obsolete/insufficient. Moreover, th ey may adversely affect the educational \nparadigm with the way current assignments and evalu ations are performed. On the other", "doc_id": "bfa53f84-e81b-4d85-80a8-28e0864774aa", "embedding": null, "doc_hash": "771f0d8053019949b356542730efc955873b26aa7fd13256d820ac0669daee3d", "extra_info": null, "node_info": {"start": 3264, "end": 7279}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "45838104-6e6e-4c29-beda-59d5941d912e", "3": "9015c73c-a129-4422-bb81-2ad60b3e0458"}}, "__type__": "1"}, "9015c73c-a129-4422-bb81-2ad60b3e0458": {"__data__": {"text": "deep learning \narchitecture) [7], robustly optimized BERT pretrain ing approach (RoBERTa) [8], and text-to-text \ntransfer transformer (T5) [9]. These AI language mo dels are all based on the transformer \narchitecture and have achieved impressive results i n various natural language processing tasks. \nHowever, each model has its own strengths and weakn esses, and the most appropriate model \ndepends on the specific application and the data av ailable. \nThese powerful language models represent a technolo gical disruption to the current \nacademic and industrial landscape. They may render many existing technologies (e.g., traditional \nsearch engines) obsolete/insufficient. Moreover, th ey may adversely affect the educational \nparadigm with the way current assignments and evalu ations are performed. On the other hand, \nit may open further avenues for exploration and lea rning if used properly. In this paper, we \ndiscuss the effects that the introduction of sophis ticated have on education, software \nengineering, healthcare, and marketing. The applica tions, drawbacks, and possible research \ndirections are presented in the next few sections. However, the effects of the latest Chatbot \nlanguage models are still being felt and more appli cations/drawbacks are coming up every day. \nThe remainder of this paper is organized as follows . In section 2 we present the possible \napplications of ChatGPT in the four identified fiel ds. The drawbacks are discussed in section 3. \nFuture research directions are explored in section 4. We conclude in section 5. \n \n2 Applications of ChatGPT and Language Models \n In this section, we highlight the possible applica tions of ChatGPT, as well as other \nadvanced language models being rolled out, in the f our aforementioned fields. As more people \nuse and adopt these AI tools, other avenues are pos sible and this is an ongoing and evolving \ntopic. \n2.1 Education \n Language models has several applications in educati on, such as providing personalized \nlearning experiences, generating test questions and answers, and facilitating online discussions. \nIt can also assist teachers in grading assignments and providing feedback to students. Some of \nthe activities involved include: \n \u2022 Language models can assist in providing person alized learning experiences by \nanalyzing student performance data and generating a daptive learning pathways. It can \nrecommend appropriate learning materials, answer st udents\u2019 questions, and provide feedback \non assignments. \n \u2022 Generate test questions and answers for stude nts, which can save time for teachers \nand ensure that tests cover a range of topics and l evels of difficulty. \n \u2022 Facilitate online discussions between student s and teachers by generating \nconversation prompts, answering questions, and prov iding feedback on responses. This can \nenhance collaboration and engagement in online lear ning environments. \n \u2022 Assist teachers in grading assignments and pr oviding feedback to students. It can \nidentify areas where students need improvement and suggest ways to improve their work. \n \u2022 Assist language learners by generating exerci ses, providing pronunciation feedback, \nand answering questions about grammar and vocabular y. It can also provide conversational \npractice for language learners by generating dialog ue prompts. \n \u2022 Assist special education students by generati ng alternative learning materials, \nproviding additional explanations, and answering qu estions in a way that is tailored to their \nindividual needs. \n \nOverall, ChatGPT\u2019s education applications have the potential to enhance student learning \nexperiences, provide teachers with valuable resourc es and assistance, and improve the efficiency \nand effectiveness of online learning environments. Tapalova and Zhiyenbayeva [10] recently \nexplored the possibilities of AI in education. A su rvey of educators at their institute indicated that \neducation can be made more effective with the help of AI. More specifically, AI can facilitate \npersonalization of the educational activities,", "doc_id": "9015c73c-a129-4422-bb81-2ad60b3e0458", "embedding": null, "doc_hash": "da74f35db0f361856268c4971ee62c81b41a2a76b5dc905de3fc44d9f85e6e74", "extra_info": null, "node_info": {"start": 7211, "end": 11311}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "bfa53f84-e81b-4d85-80a8-28e0864774aa", "3": "3e0c2955-89ea-4a88-ab65-323f73c022a6"}}, "__type__": "1"}, "3e0c2955-89ea-4a88-ab65-323f73c022a6": {"__data__": {"text": "\npractice for language learners by generating dialog ue prompts. \n \u2022 Assist special education students by generati ng alternative learning materials, \nproviding additional explanations, and answering qu estions in a way that is tailored to their \nindividual needs. \n \nOverall, ChatGPT\u2019s education applications have the potential to enhance student learning \nexperiences, provide teachers with valuable resourc es and assistance, and improve the efficiency \nand effectiveness of online learning environments. Tapalova and Zhiyenbayeva [10] recently \nexplored the possibilities of AI in education. A su rvey of educators at their institute indicated that \neducation can be made more effective with the help of AI. More specifically, AI can facilitate \npersonalization of the educational activities, incr ease availability of resources, improve \nadaptability of the educational material to individ ual student needs, provide prompt and \ncontinuous feedback, and improve mental motivation and stimulation. However, it is essential to \nensure that ChatGPT is used responsibly and thought fully, with considerations for potential \nbiases and ethical concerns. In another study, Kuma r and Boulanger [11] explored the use of deep \nlearning AI to automatically grade essays using rub ric instead of holistic scores. They concluded \nthat it is possible for language models to aid stud ents in learning proper writing and its strategies. \nLanguage models can be a useful tool for teaching an d learning, particularly in the field of \nlanguage arts and writing. Here are a few ways that language models can be used to enhance \nteaching: \n \u2022 Writing prompts: Language models can be used t o spark students\u2019 creativity and \nengage them in writing. For example, enter a topic or theme and ask the Chabot to generate a \nwriting prompt for students to work on. \n \u2022 Writing feedback: After students have written a piece, the language model can \nprovide feedback on their work. For example, ChatGP T can analyze the writing for grammar, \npunctuation, and spelling errors, as well as provid e suggestions for improving the overall \nstructure and flow of the writing. \n \u2022 Language practice: Advanced language models ca n help students practice their \nlanguage skills. For example, one can ask ChatGPT t o provide synonyms or antonyms for certain \nwords, or to provide sample sentences using certain grammar structures or vocabulary. \n \u2022 Research assistance: language models can be u sed to assist students in their \nresearch by acting as an advanced search engine. Ho wever, this disrupts the current \nhomework/assignment models with the lack of ability to detect plagiarism using ChatGPT. \n \u2022 Personalized learning: Language models can als o be used to create personalized \nlearning experiences for students. For example, Cha tGPT can be used to provide feedback and \nguidance to individual students based on their spec ific strengths and weaknesses in writing. \n \nHowever, assessing the impact of using ChatGPT in t eaching and learning is important to \ndetermine whether it is an effective tool for impro ving students\u2019 skills. The following are a few \nways to measure the impact of using ChatGPT in teac hing: \n \u2022 Pre- and post-assessments to measure the impr ovement in students\u2019 skills. The \nassessments should be aligned with the learning obj ectives and outcomes of using ChatGPT. For \nexample, assess students\u2019 writing skills before and after using ChatGPT to see if there is an \nimprovement in the quality of their writing. \n \u2022 Analyzing student work to see if there is an improvement in their skills. A rubric to \nassess students\u2019 writing, and compare their work be fore and after using ChatGPT to see if there \nis an improvement in areas such as grammar, sentenc e structure, and vocabulary. \n \u2022 Student feedback. Surveys or focus groups can be used to gather their feedback on \nthe usefulness of ChatGPT in improving their skills , as well as their overall experience of using", "doc_id": "3e0c2955-89ea-4a88-ab65-323f73c022a6", "embedding": null, "doc_hash": "82f5548dc3e33700854bfc21389ebbcfb153f20d4c08ba5ec87a45fe2d8d66fc", "extra_info": null, "node_info": {"start": 11329, "end": 15345}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "9015c73c-a129-4422-bb81-2ad60b3e0458", "3": "eb6f75f5-4119-4a47-80f1-f967c1299210"}}, "__type__": "1"}, "eb6f75f5-4119-4a47-80f1-f967c1299210": {"__data__": {"text": "impr ovement in students\u2019 skills. The \nassessments should be aligned with the learning obj ectives and outcomes of using ChatGPT. For \nexample, assess students\u2019 writing skills before and after using ChatGPT to see if there is an \nimprovement in the quality of their writing. \n \u2022 Analyzing student work to see if there is an improvement in their skills. A rubric to \nassess students\u2019 writing, and compare their work be fore and after using ChatGPT to see if there \nis an improvement in areas such as grammar, sentenc e structure, and vocabulary. \n \u2022 Student feedback. Surveys or focus groups can be used to gather their feedback on \nthe usefulness of ChatGPT in improving their skills , as well as their overall experience of using \nthe tool. \n \u2022 Observations to assess student\u2019s engagement a nd level of participation. This involve \nobserving their interactions with ChatGPT, their le vel of motivation, and their ability to use the \ntool effectively. \n \u2022 Comparison with control group. If possible, t he progress of students who used \nChatGPT can be compared with a control group of stu dents who did not use the tool. This can \nhelp to isolate the impact of using ChatGPT and det ermine whether it was a significant factor in \nimproving students\u2019 skills. \n \nBy using these methods to assess the impact of usin g language models in teaching, it is \npossible to determine whether it is an effective to ol for improving students\u2019 skills, and make any \nnecessary adjustments to the teaching methods to fu rther enhance learning outcomes. \n \n2.2 Software Engineering \n ChatGPT can be used in software engineering for ta sks such as generating code, \ndebugging, and software testing. It can also help d evelopers in natural language processing tasks, \nsuch as analyzing user requirements, and generating user interfaces. This can be accomplished \nas follows: \n \u2022 Code generation: Generate code snippets for d evelopers based on natural language \ndescriptions of the desired functionality. This can save time and improve efficiency in the \nsoftware development process. \n \u2022 Debugging: Assist in debugging code by identi fying errors and suggesting fixes based \non natural language descriptions of the issue. \n \u2022 Software testing: generate test cases and tes t data based on natural language \ndescriptions of the desired test scenarios. This ca n improve the efficiency and effectiveness of \nsoftware testing. \n \u2022 Natural language processing: Assist developer s in natural language processing tasks, \nsuch as analyzing user requirements, generating use r interfaces, and providing Chatbot \ninteractions with users. \n \u2022 Documentation generation: Generate software d ocumentation based on natural \nlanguage descriptions of the software\u2019s functionali ty. This can save time for developers and \nimprove the quality and completeness of the documen tation. \n \u2022 Collaboration and knowledge sharing: Facilita te collaboration and knowledge \nsharing between developers by generating conversati on prompts, answering questions, and \nproviding feedback on responses. This can enhance c ommunication and efficiency in software \ndevelopment teams. \n \nOverall, ChatGPT\u2019s software engineering application s have the potential to improve \nefficiency and effectiveness in the software develo pment process, enhance collaboration and \nknowledge sharing, and improve the quality of softw are documentation. Raychev et al. [12] \nidentified early on the potential of natural langua ge processing in synthesizing code completion \nand predicting the probability of sentences. In add ition, they used a similar approach to predict \nsyntactic and semantic variable types and identifie d names. Such efforts have resulted in a slew \nof studies that research the role of language model s in specific software engineering problems \n[13]. In an another study, Tu et al. [14] investiga ted the role of language models in predicting the \nrepetitive, regular, and typical code snippets in h uman-written programs. This has the potential", "doc_id": "eb6f75f5-4119-4a47-80f1-f967c1299210", "embedding": null, "doc_hash": "5b3f3ae0619668d33d2d15218851766f9bf75404fa0e9bca108f75ea1287db27", "extra_info": null, "node_info": {"start": 15425, "end": 19472}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "3e0c2955-89ea-4a88-ab65-323f73c022a6", "3": "e8b520f8-96ef-4abe-b6f5-fcb26be295e7"}}, "__type__": "1"}, "e8b520f8-96ef-4abe-b6f5-fcb26be295e7": {"__data__": {"text": "application s have the potential to improve \nefficiency and effectiveness in the software develo pment process, enhance collaboration and \nknowledge sharing, and improve the quality of softw are documentation. Raychev et al. [12] \nidentified early on the potential of natural langua ge processing in synthesizing code completion \nand predicting the probability of sentences. In add ition, they used a similar approach to predict \nsyntactic and semantic variable types and identifie d names. Such efforts have resulted in a slew \nof studies that research the role of language model s in specific software engineering problems \n[13]. In an another study, Tu et al. [14] investiga ted the role of language models in predicting the \nrepetitive, regular, and typical code snippets in h uman-written programs. This has the potential \nto improve code suggestions in automatic completion systems. Furthermore, Allamanis et al. [15] \ninvestigated the role of language models in detecti ng software bugs [16]. In another avenue, \nlanguage models have been shown to useful in the de velopment and update of software \ncomments and documentation. These efforts show that it is possible to tap into the power of \nlanguage models in the software engineering domain. \n \n2.3 Healthcare \n ChatGPT can assist healthcare professionals by pro viding patient triage, symptom \nanalysis, and medical diagnosis. It can also aid in drug discovery and clinical trials. This can be \naccomplished as follows: \n \u2022 Patient Care: Assist healthcare professionals by providing personalized care to \npatients. It can answer questions about medical con ditions, treatments, and medications, and \nprovide recommendations based on patient symptoms a nd medical history. \n \u2022 Electronic health records (EHR): Assist in up dating electronic health records by \nanalyzing natural language descriptions of patient conditions and treatments and generating \ncorresponding entries in the EHR. \n \u2022 Medical education: Assist medical students an d professionals by generating medical \ncase studies, answering questions about medical ter minology, and providing educational \nresources for medical training. \n \u2022 Mental health: Assist in mental health care b y providing personalized support and \nresources to patients. It can answer questions abou t mental health conditions, provide coping \nstrategies and relaxation techniques, and offer sup port for patients experiencing mental health \ncrises. \n \u2022 Clinical trials: assist in clinical trials by generating eligibility criteria, screening \nquestions, and informed consent forms based on natu ral language descriptions of the trial\u2019s \nobjectives. \n \u2022 Telemedicine: Assist in telemedicine by facil itating communication between \nhealthcare professionals and patients. It can answe r questions about telemedicine procedures, \nprovide technical support for patients using teleme dicine tools, and assist in scheduling \nappointments. \n \nIn general, ChatGPT\u2019s healthcare applications have the potential to improve patient care, \nenhance medical education and training, and improve efficiency in healthcare operations. Adlung \net al. [17v] identified two challenges facing artif icial intelligence in clinical decision making, \nmainly explainability and causability. Explainabili ty refers to the ability of the model to provide \nresults that can be justified (e.g., significant fa ctors that have statistical association with the \noutput). Although examples on the Web were able to show ChatGPT giving reasons for their \nanswers or the lack of a correct answer, language m odels still has the potential to open up deep \nlearning models for easy explanations and transpare ncy. This also relates to causability, from a \nlegal and regulatory perspective, adopting artifici al intelligence algorithms in clinical decision \nmaking may require such methods to provide clear ex planations on why a certain output was \ngenerated. Moreover, Wu et al. [18] pointed out the importance of literature review in the field \nof regulatory science and the role that language mo dels can play in", "doc_id": "e8b520f8-96ef-4abe-b6f5-fcb26be295e7", "embedding": null, "doc_hash": "41ac1a95af4be8d599f5220fd59befdebe356d9d3950212fbd7de6d1a921cc6c", "extra_info": null, "node_info": {"start": 19379, "end": 23486}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "eb6f75f5-4119-4a47-80f1-f967c1299210", "3": "041abddf-0d22-4833-a7aa-45330b5bb49e"}}, "__type__": "1"}, "041abddf-0d22-4833-a7aa-45330b5bb49e": {"__data__": {"text": "Explainabili ty refers to the ability of the model to provide \nresults that can be justified (e.g., significant fa ctors that have statistical association with the \noutput). Although examples on the Web were able to show ChatGPT giving reasons for their \nanswers or the lack of a correct answer, language m odels still has the potential to open up deep \nlearning models for easy explanations and transpare ncy. This also relates to causability, from a \nlegal and regulatory perspective, adopting artifici al intelligence algorithms in clinical decision \nmaking may require such methods to provide clear ex planations on why a certain output was \ngenerated. Moreover, Wu et al. [18] pointed out the importance of literature review in the field \nof regulatory science and the role that language mo dels can play in accelerating this review. \nThese factors bring language models to the fore of the medical AI domain. Lederman [19] \ndiscussed clinical natural language processing (cNLP ) that use textual data in health records to \nsupport the clinical decision making process. They argued for a rethink of cNLP systems to \nimprove their practical adaptation and deployment [ 20]. Specifically, they identify several factors \nthat hinder the usability of the cNLP systems, which may be overcome by advanced language \nmodels, including handling of complex language proc essing, ability to answer \u201chow\u201d and \u201cwhy\u201d \nquestions, and the problem of explainability. In an interesting study, Liu et al. [21] investigated \nthe role of language models in the development and discovery of drugs. Mainly, language models \ncan have a great role in the quick identification o f targets, optimization of clinical trials, \nfacilitation of decision making from a regulatory p erspective, and promoting pharmacovigilance \n[22]. Bhatnagar et al. [23] also reviewed the role of language models in discovering new drugs, \nclinical trials, and pharmacovigilance. \nThe topic of pharmacovigilance is another interesti ng are, where language models can \ngauge, analyze, and detect drug-related problems or adverse effects/interactions from users\u2019 \nprompts and discussions [24]. Ball and Pan [25] exp lored the use of language models in the \nprocessing of the individual case safety reports su bmitted to the Food and Drug Administration \nas part of their adverse event reporting system. Th ey identified several problems that need to be \nresolved in order to facilitate the acceptance of l anguage models in pharmacovigilance. Koneti \net al. [26] proposed using language models in drug development by extracting unstructured data \nfrom pharmacokinetics and pharmacodynamics study re ports. Several language models were \nproposed recently for the purpose of medical text m ining. Wang et al. [27] developed the \n\u201cDeepCausality\u201d model, which is able to include AI language models in order to create a causal \ninference model from fee text. They demonstrated it s effectiveness in detecting idiosyncratic \ndrug-induced liver injury with high accuracy. Lee et al. [28] proposed bidirectional encoder \nrepresentations from transformers for biomedical te xt mining (BioBert) model, which was able \nto answer biomedical questions, extract relations, and recognize named biomedical entities with \nimproved accuracy. Similarly, ClinicalBert [29,30] was proposed to predict hospital readmissions \nusing medical text data from hospital admission not es and discharge summaries. \n \n2.4 Marketing \n ChatGPT can be used in marketing to generate produ ct descriptions, customer reviews, \nand social media content. It can also assist in Cha tbot interactions with customers and provide \npersonalized recommendations based on customer pref erences. The following are some \nelaborations on ChatGPT\u2019s marketing applications: \n \n \u2022 Content creation: Assist in content creation for marketing campaigns by generating \nideas for social media posts, email marketing campa igns, and blog articles. It can also generate \nheadlines, product descriptions, and promotional me ssages based on natural language \ndescriptions of the marketing objectives.", "doc_id": "041abddf-0d22-4833-a7aa-45330b5bb49e", "embedding": null, "doc_hash": "b589fbc40da9cbad2ac9a29578a0dd17adb08a7f41cd8f23434f4cce232c9d85", "extra_info": null, "node_info": {"start": 23506, "end": 27610}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "e8b520f8-96ef-4abe-b6f5-fcb26be295e7", "3": "eec8c5c1-025d-4d48-8c95-02960991813f"}}, "__type__": "1"}, "eec8c5c1-025d-4d48-8c95-02960991813f": {"__data__": {"text": "was proposed to predict hospital readmissions \nusing medical text data from hospital admission not es and discharge summaries. \n \n2.4 Marketing \n ChatGPT can be used in marketing to generate produ ct descriptions, customer reviews, \nand social media content. It can also assist in Cha tbot interactions with customers and provide \npersonalized recommendations based on customer pref erences. The following are some \nelaborations on ChatGPT\u2019s marketing applications: \n \n \u2022 Content creation: Assist in content creation for marketing campaigns by generating \nideas for social media posts, email marketing campa igns, and blog articles. It can also generate \nheadlines, product descriptions, and promotional me ssages based on natural language \ndescriptions of the marketing objectives. \n \n \u2022 Customer service: Assist in customer service by providing personalized support to \ncustomers. It can answer questions about products a nd services, provide recommendations \nbased on customer preferences, and assist in resolv ing customer issues and complaints. \n \n \u2022 Lead generation: Assist in lead generation by analyzing customer data and \ngenerating natural language descriptions of potenti al customers. It can also assist in lead \nqualification by analyzing customer responses and i dentifying potential leads. \n \n \u2022 Market research: Assist in market research by generating surveys, analyzing \ncustomer feedback, and identifying trends and insig hts based on natural language descriptions \nof the research objectives. \n \n \u2022 Personalization: Assist in personalizing mark eting campaigns by analyzing customer \ndata and generating personalized recommendations fo r products and services. It can also assist \nin tailoring marketing messages to specific custome r segments based on natural language \ndescriptions of the customer demographics and prefe rences. \n \n \u2022 Sales: Assist in sales by generating personal ized product recommendations and \nassisting in the sales process. It can also assist in upselling and cross-selling by analyzing \ncustomer data and generating natural language descr iptions of potential add-ons or upgrades. \n \nChatGPT\u2019s marketing applications have the potential to improve efficiency and \neffectiveness in marketing campaigns, enhance custo mer experience and satisfaction, and \nimprove sales performance. Verma et al. [31] discus sed the role of recent disruptive \ntechnologies, mainly AI, in business operations. On e of the areas that they have identified was \nthe use of Chatbots and language models to improve customer experience [32] and customer \nrelationship management (CRM) systems. Language mode ls and Chatbots offer great advantages \nin the form of faster and automated access to data, simpler and efficient processes, accuracy, \nand cost effectiveness [33]. Similarly, De Mauro et al. [34] published a recent taxonomy of the \nuse of machine learning and AI in marketing. The au thors have identified several use cases of AI \nin marketing and divided those into customer side v ersus business side. On the customer facing \nside, they identified personalization of offers, co mmunication, recommendations, and \nassortments as candidates for improvements. Moreove r, they indicated that the consumption \nexperience can also be improved via experience impr ovement and digital customer service. On \nthe business side, machine learning can be benefici al in market understanding and customer \nsensing, among other avenues. In a recent literatur e review of marketing and AI, Duarte et al. \n[35] have identified recommender systems and text a nalysis as promising areas of useful \nChatbots usage in marketing. De Bruyn et al. [36] i nvestigated the opportunities and pitfalls of \nusing AI in marketing. They have identified several risks associated with adopting new AI \ndisruptive technologies, mainly bias, explainabilit y, control, and unsafe/unrealistic learning \nenvironments. Moreover, they conclude with a warnin g of the possibility of AI failure in this \ndomain if these challenges are not resolved by the", "doc_id": "eec8c5c1-025d-4d48-8c95-02960991813f", "embedding": null, "doc_hash": "9605d2c8f2cc0c372b29e7e113b7a25c2236950a29e295e3bad68e986ab8e169", "extra_info": null, "node_info": {"start": 27622, "end": 31695}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "041abddf-0d22-4833-a7aa-45330b5bb49e", "3": "5c196df3-e04a-46bc-89a1-d3406c2f0357"}}, "__type__": "1"}, "5c196df3-e04a-46bc-89a1-d3406c2f0357": {"__data__": {"text": "via experience impr ovement and digital customer service. On \nthe business side, machine learning can be benefici al in market understanding and customer \nsensing, among other avenues. In a recent literatur e review of marketing and AI, Duarte et al. \n[35] have identified recommender systems and text a nalysis as promising areas of useful \nChatbots usage in marketing. De Bruyn et al. [36] i nvestigated the opportunities and pitfalls of \nusing AI in marketing. They have identified several risks associated with adopting new AI \ndisruptive technologies, mainly bias, explainabilit y, control, and unsafe/unrealistic learning \nenvironments. Moreover, they conclude with a warnin g of the possibility of AI failure in this \ndomain if these challenges are not resolved by the implicit marketing knowledge transfer to AI \nmodels. \n \n3 Drawbacks of Language Models \n Although language models and Chatbots clearly offe r great opportunities, they have \ninherent shortcomings that limit their applicabilit y, adoption, and usefulness. In the next few \nsubsections, we go through these drawbacks in detai l. \n3.1 Bias \n Language models can exhibit bias if the training da ta used to create them is biased. As \nSchramowski et al. [37] pointed out, large pre-trai ned models that try to mimic natural languages, \nmay end up repeating the same unfairness and prejud ices. This can lead to discriminatory or \ninaccurate analyses and recommendations. Moreover, this may lead to public outcry (i.e., \npolitical, social, and legal) against the commercia l applications. These biases manifests \nthemselves in several ways, as follows: \n \u2022 Training data bias: Language models are typica lly trained on large datasets of \nhuman language. If these datasets are biased in som e way (e.g., based on race, gender, \nsocioeconomic status, etc.), then the model may lea rn and replicate these biases in its \nresponses. For example, if the training data is bia sed towards a particular gender, then the \nmodel may generate responses that are more favorabl e towards that gender. \n \u2022 User interaction bias: The responses generate d by Chatbots are based on the input \nthey receive from users. If users consistently ask biased or prejudiced questions, then the \nmodel may learn and replicate these biases in its r esponses. For example, if users frequently ask \nquestions that are discriminatory towards a particu lar group, then the model may generate \nresponses that perpetuate these biases. \n \u2022 Algorithmic bias: The algorithms used to trai n and operate language models and \nChatbots may also introduce biases. For example, if the model is trained to optimize for a \nparticular metric (e.g., accuracy, engagement, etc. ), then it may prioritize generating responses \nthat optimize for that metric, even if those respon ses are biased in some way. \n \u2022 Contextual bias: Chatbots generate responses based on the context they receive \nfrom users. If the context is biased in some way (e .g., based on the user\u2019s location, language, \netc.), then the model may generate biased responses . For example, if a user is asking questions \nabout a particular culture or religion, and the mod el is not trained on that culture or religion, it \nmay generate biased responses due to its lack of kn owledge. \n It is important to note that bias in language mode ls are not necessarily intentional or \nmalicious. Although this sometimes is hard to prove or justify to the non-technical public. \nMoreover, it can have harmful consequences, such as perpetuating stereotypes, reinforcing \ndiscriminatory attitudes, or excluding certain grou ps from access to information and resources. \nTo mitigate these risks, it is of paramount importa nce to train and operate the models in a \nresponsible and ethical manner, by carefully select ing and monitoring training data, \nincorporating diversity and inclusion consideration s, and regularly auditing the", "doc_id": "5c196df3-e04a-46bc-89a1-d3406c2f0357", "embedding": null, "doc_hash": "c954a2aab1e96c45105df6cb42564aa1a918c35259b2f21093ed6095e0a43d32", "extra_info": null, "node_info": {"start": 31709, "end": 35641}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "eec8c5c1-025d-4d48-8c95-02960991813f", "3": "76d2b8ea-8fc5-4e16-8413-1e632581c4d7"}}, "__type__": "1"}, "76d2b8ea-8fc5-4e16-8413-1e632581c4d7": {"__data__": {"text": "if a user is asking questions \nabout a particular culture or religion, and the mod el is not trained on that culture or religion, it \nmay generate biased responses due to its lack of kn owledge. \n It is important to note that bias in language mode ls are not necessarily intentional or \nmalicious. Although this sometimes is hard to prove or justify to the non-technical public. \nMoreover, it can have harmful consequences, such as perpetuating stereotypes, reinforcing \ndiscriminatory attitudes, or excluding certain grou ps from access to information and resources. \nTo mitigate these risks, it is of paramount importa nce to train and operate the models in a \nresponsible and ethical manner, by carefully select ing and monitoring training data, \nincorporating diversity and inclusion consideration s, and regularly auditing the model for \npotential biases. \n \n3.2 Lack of Transparency \n Language models, and deep learning models in genera l, are called \u201cblack box\u201d models as \ntheir results can be difficult to interpret and und erstand, making it challenging for researchers to \nassess their validity and accuracy [38]. These mode ls lack transparency, meaning that it is often \ndifficult to understand how the model arrived at a particular output or decision. This lack of \ntransparency can be problematic for several reasons : \n 1. Debugging: If the model generates unexpecte d or incorrect output, it can be \nchallenging to identify the source of the problem w ithout understanding how the model arrived \nat its decision. \n 2. Accountability: In some cases, the output g enerated by the model may have \nsignificant consequences for individuals or society as a whole (e.g., in healthcare or criminal \njustice). If the model lacks transparency, it can b e difficult to hold it accountable for its \ndecisions. \n 3. Bias: As mentioned earlier, language models can be biased in various ways, such as \nin the training data or algorithms used. Without tr ansparency, it can be difficult to identify and \ncorrect these biases. \n 4. Trust: In many cases, users may be hesitant to trust the output generated by the \nmodel if they don\u2019t understand how it arrived at it s decision. This have ramifications in \nobtaining regulatory approvals and adoption by the public. \n \n \n3.3 Explainability \n Researchers are developing new techniques for maki ng deep learning models more \ninterpretable and explainable. For example, techniq ues such as attention mechanisms [39] or \nsaliency maps can highlight which parts of the inpu t the model is focusing on to make its decision \n[40]. Hicks et al. [41] argued that deep learning p redictions and decisions need to be \naccompanied by explanations so that the doctors res ponsible for the clinical decision-making \nprocess trust, understand, and validate such models [42-44]. To this end, they proposed a new \nmethod called electrocardiogram gradient class acti vation map, which produces explanations for \nthe results of the electrocardiogram (EEG) analysis . Along the same line of EEG analysis, \nKhasawneh et al. [45,46] proposed treating the signa l in a similar fashion to the clinicians and \nperform the signal inspection visually using deep l earning object detection algorithms. In the \ncontext of language models, further explainability can be achieved via the following practices: \n \u2022 Documentation: Developers can document how th e model was trained, what data \nwas used, what decisions were made in the training process, and what assumptions were \nmade. Moreover, ethical standards can be developed to standardize the training process. This \ncan help increase transparency and accountability. \n \u2022 Auditing: Regular auditing of the model can h elp identify and correct biases, as well \nas provide insights into how the model is making de cisions. \n \u2022 Collaboration: Collaboration between develope rs, users, and experts in relevant \nfields can help increase transparency and ensure th at the", "doc_id": "76d2b8ea-8fc5-4e16-8413-1e632581c4d7", "embedding": null, "doc_hash": "36d339883c37d21f760fee15228d35e8c8b938f3ee9c6267aab5be6265b342d8", "extra_info": null, "node_info": {"start": 35598, "end": 39579}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "5c196df3-e04a-46bc-89a1-d3406c2f0357", "3": "6695aaa4-3f68-4863-9eff-8e37406349e8"}}, "__type__": "1"}, "6695aaa4-3f68-4863-9eff-8e37406349e8": {"__data__": {"text": "the signal inspection visually using deep l earning object detection algorithms. In the \ncontext of language models, further explainability can be achieved via the following practices: \n \u2022 Documentation: Developers can document how th e model was trained, what data \nwas used, what decisions were made in the training process, and what assumptions were \nmade. Moreover, ethical standards can be developed to standardize the training process. This \ncan help increase transparency and accountability. \n \u2022 Auditing: Regular auditing of the model can h elp identify and correct biases, as well \nas provide insights into how the model is making de cisions. \n \u2022 Collaboration: Collaboration between develope rs, users, and experts in relevant \nfields can help increase transparency and ensure th at the model is being used in an ethical and \nresponsible manner. \n While these approaches can help address the lack o f transparency in deep learning, it is \nimportant to acknowledge that achieving full transp arency may not be possible or desirable (e.g., \nproprietary or copyrighted/patented models) in all cases. \n \n3.4 Over-reliance \n Researchers, professionals, or students may become over-reliant on Chatbots, language \nmodels, and AI in general. Thus, they may neglect c ritical thinking, leading to errors and \ninaccuracies in their research, studies, or practic e/work. Such over-reliance can happen in several \nways, as in the following examples: \n \u2022 Dataset selection: Researchers may rely too h eavily on Chatbots to generate \nsynthetic data or to augment existing datasets. Thi s can be problematic if the generated data is \nbiased or does not accurately reflect the real-worl d data. \n \u2022 Hypothesis generation: Language models can gen erate hypotheses or research \nquestions based on input from researchers. While th is can be a useful tool for exploring new \nareas of research, researchers should be cautious n ot to rely too heavily on the model\u2019s \nsuggestions without independent validation. \n \u2022 Data analysis: Chatbots can be used to analyz e and summarize large datasets. While \nthis can save time and resources, researchers shoul d be cautious not to rely too heavily on the \nmodel\u2019s output without independent verification. \n \u2022 Model selection: Researchers may choose to us e ChatGPT (or other AI language \nmodels) as their primary research tool, rather than as one tool among many. This can lead to \nover-reliance on the model\u2019s output and a failure t o consider alternative hypotheses or \nmethods. \n \nOver-reliance can lead to several problems, includi ng: \n \u2022 Biases: As we discussed earlier, Chatbots and language models can be biased in \nvarious ways. If researchers rely too heavily on th e model\u2019s output, they may unknowingly \nreplicate or amplify these biases. \n \u2022 Errors: ChatGPT (like all models) is not infa llible. If researchers rely too heavily on \nthe model\u2019s output, they may introduce errors or in accuracies into their research. \n \u2022 Over-generalization: ChatGPT is trained on a large corpus of text and may not \naccurately reflect the nuances and complexities of the real world. If researchers rely too heavily \non the model\u2019s output, they may over-generalize or oversimplify their findings. \n \nTo avoid over-reliance on AI and language models, r esearchers should be cautious in their \nuse of the model and should use it in conjunction w ith other research methods and tools. They \nshould also be aware of the model\u2019s limitations and potential biases, and should take steps to \nmitigate these risks. \n \n3.5 Ethical Concerns \n ChatGPT can raise ethical concerns such as privacy violations and job displacement (i.e., \ninvoluntary job loss). ChatGPT can generate respons es that", "doc_id": "6695aaa4-3f68-4863-9eff-8e37406349e8", "embedding": null, "doc_hash": "e133cabca104cddd8f0adf1f7a9ad54f360b39d4acb42fdcc1f6ab5d7317ca46", "extra_info": null, "node_info": {"start": 39607, "end": 43368}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "76d2b8ea-8fc5-4e16-8413-1e632581c4d7", "3": "f7c1761b-4d8d-41a2-85eb-b597ef039513"}}, "__type__": "1"}, "f7c1761b-4d8d-41a2-85eb-b597ef039513": {"__data__": {"text": "ChatGPT is trained on a large corpus of text and may not \naccurately reflect the nuances and complexities of the real world. If researchers rely too heavily \non the model\u2019s output, they may over-generalize or oversimplify their findings. \n \nTo avoid over-reliance on AI and language models, r esearchers should be cautious in their \nuse of the model and should use it in conjunction w ith other research methods and tools. They \nshould also be aware of the model\u2019s limitations and potential biases, and should take steps to \nmitigate these risks. \n \n3.5 Ethical Concerns \n ChatGPT can raise ethical concerns such as privacy violations and job displacement (i.e., \ninvoluntary job loss). ChatGPT can generate respons es that may violate users\u2019 privacy, and the \nuse of ChatGPTs in various industries may lead to j ob displacement. There are several ethical \nconcerns associated with ChatGPT and other AI model s. Here are a few examples: \n \u2022 Bias: As we discussed earlier, language model s can be biased in various ways, such \nas in the training data or algorithms used. These b iases can lead to unfair or discriminatory \noutcomes, such as in employment, healthcare, or cri minal justice. \n \u2022 Privacy: Chatbots can generate highly persona lized output based on input from \nusers, which can raise privacy concerns. For exampl e, if a user inputs sensitive information into \nthe model (such as health or financial data), the m odel\u2019s output could reveal that information \nto others. \n \u2022 Accountability: ChatGPT (and other AI models) can make decisions or generate \noutput with significant consequences for individual s or society as a whole (e.g., in healthcare or \ncriminal justice). If the model makes an incorrect or biased decision, it can be difficult to hold it \naccountable for its actions. \n \u2022 Transparency: As we discussed earlier, deep l earning can lack transparency, \nmeaning that it is often difficult to understand ho w the model arrived at a particular output or \ndecision. This lack of transparency can make it dif ficult to identify and correct biases or to hold \nthe model accountable for its actions. \n \u2022 Misuse: ChatGPT can be misused for nefarious purposes, such as generating fake \nnews or propaganda. Moreover, academia is ringing t he alarm bills about the great possibilities \nfor cheating on academic assignments using language models and Chatbots, coupled with the \nlagging behind of cheating detections software on t his problem. This can have serious \nconsequences for individuals and society as a whole . \n \nTo address these ethical concerns, researchers, dev elopers, and users of ChatGPT should \nprioritize ethical considerations throughout the mo del\u2019s development and use. This can include: \n \u2022 Fairness: Ensuring that the model is trained on diverse and representative data, and \nthat it does not unfairly discriminate against any particular group of people. \n \u2022 Privacy: Ensuring that the model is used in a way that respects users\u2019 privacy and \nthat sensitive data is protected. \n \u2022 Accountability: Ensuring that there are mecha nisms in place to hold the model \naccountable for its decisions and actions. \n \u2022 Transparency: Ensuring that the model\u2019s outpu t is transparent and explainable, so \nthat users can understand how the model arrived at its decisions. \n \n \u2022 Responsible use: Ensuring that the model is u sed in an ethical and responsible \nmanner, and that it is not misused for nefarious pu rposes. \n Each one of these considerations open several aven ues for research. For example, the \nalgorithms for calculating similarity scores and ch eating detection need to be developed to take \nunder consideration the availability of powerful Ch atbots like ChatGPT. \n4 Future Research Directions \n \n4.1 Explainability", "doc_id": "f7c1761b-4d8d-41a2-85eb-b597ef039513", "embedding": null, "doc_hash": "feb29eb1299f6ccd83e275f91d4880cff76617d58c247b2c4d71c102282e4edd", "extra_info": null, "node_info": {"start": 43437, "end": 47247}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "6695aaa4-3f68-4863-9eff-8e37406349e8", "3": "0780cde1-9a19-4cbd-9525-12a03b247a78"}}, "__type__": "1"}, "0780cde1-9a19-4cbd-9525-12a03b247a78": {"__data__": {"text": " \u2022 Accountability: Ensuring that there are mecha nisms in place to hold the model \naccountable for its decisions and actions. \n \u2022 Transparency: Ensuring that the model\u2019s outpu t is transparent and explainable, so \nthat users can understand how the model arrived at its decisions. \n \n \u2022 Responsible use: Ensuring that the model is u sed in an ethical and responsible \nmanner, and that it is not misused for nefarious pu rposes. \n Each one of these considerations open several aven ues for research. For example, the \nalgorithms for calculating similarity scores and ch eating detection need to be developed to take \nunder consideration the availability of powerful Ch atbots like ChatGPT. \n4 Future Research Directions \n \n4.1 Explainability \n One of the most crucial research directions is to develop methods to make deep learning \nin general and language models in particular more e xplainable. Explainability refers to the ability \nto understand how a model arrived at its output or decision, and to be able to explain that \nprocess in a way that is understandable to humans. This will enable researchers to understand \nthe logic behind the models\u2019 decisions and provide transparency in their output. Explainability is \nan important research direction for Chatbots, as it can help address concerns around \ntransparency and accountability. Failing to address explainability has great ramifications on the \nadoption and regulatory certification of AI techniq ues [47,48]. For example, the European general \ndata protection regulation explicitly requires deci sions made in healthcare among other areas to \nbe traceable and explainable [49]. Explainable AI ( XAI) is a research avenue in AI that gaining a \nlot of attention driven by the real-life deployment requirements of AI-based systems. A recent \nsurvey by Bai et al. [50] presents the recent advan cements toward achieving explainable AI in \npattern recognition. \nOne approach to achieving explainability is through the use of attention mechanisms. \nAttention mechanisms allow the model to focus on ce rtain parts of the input when generating its \noutput. They can generate probability distributions relating to the input, which serve as \nindicators on the importance of features. By visual izing the attention weights for each part of the \ninput [51], we can gain insight into which parts of the input the model is focusing on and how it \nis using that information to generate its output. H owever, attention mechanisms according to Liu \net al. [52] may be unable to identify the polarity of the impact of individual features due to \nsuppression effects. \nAnother approach is to use model-agnostic methods t o explain the output of deep \nlearning models. These methods aim to explain the m odel\u2019s output without needing to know the \ninternal workings of the model itself. For example, one such method is LIME (Local Interpretable \nModel-agnostic Explanations) [53], which generates a simple, interpretable model that \napproximates the behavior of the original model on a local scale. Aditya and Pal [54] proposed \nfurther refinements to LIME using Shapley values use d in game theory, which provide several \nadvantages in terms of efficiency, consistency, and symmetry. \nIn addition to these approaches, there is ongoing r esearch into developing new methods \nfor explainability. For example, recent work has ex plored the use of counterfactual explanations, \nwhich aim to explain how the output of the model wo uld have changed if certain parts of the \ninput had been different [31]. Another area of rese arch is developing methods to explain the \noutput of black box models when the input is a sequ ence of events over time, such as in the case \nof medical records or financial transactions. \nOverall, the goal of explainability research is to provide users with a better understanding \nof how language models in particular arrive at thei r output, and to provide mechanisms for \nidentifying and correcting biases or errors in the model\u2019s decision-making. This is an important \narea of research for", "doc_id": "0780cde1-9a19-4cbd-9525-12a03b247a78", "embedding": null, "doc_hash": "095f2375d10187d2c1173ea12893c2419c51167771c13debe896a2ecf9e37816", "extra_info": null, "node_info": {"start": 47235, "end": 51318}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "f7c1761b-4d8d-41a2-85eb-b597ef039513", "3": "6012fe95-dfd4-477c-b0c5-98be7f19251a"}}, "__type__": "1"}, "6012fe95-dfd4-477c-b0c5-98be7f19251a": {"__data__": {"text": " \nIn addition to these approaches, there is ongoing r esearch into developing new methods \nfor explainability. For example, recent work has ex plored the use of counterfactual explanations, \nwhich aim to explain how the output of the model wo uld have changed if certain parts of the \ninput had been different [31]. Another area of rese arch is developing methods to explain the \noutput of black box models when the input is a sequ ence of events over time, such as in the case \nof medical records or financial transactions. \nOverall, the goal of explainability research is to provide users with a better understanding \nof how language models in particular arrive at thei r output, and to provide mechanisms for \nidentifying and correcting biases or errors in the model\u2019s decision-making. This is an important \narea of research for ensuring the ethical use of Ch atGPTs in a variety of applications. \n \n4.2 Bias Detection and Mitigation \n Another research direction is to detect and mitiga te bias in ChatGPTs. Researchers can \ndevelop methods to identify and eliminate bias from Chatbots by using techniques such as \nadversarial training [55]. This method fine tunes l anguage models and deep learning models in \ngeneral through the introduction of adversarial sam ples in the training set, which tends to \nincrease the robustness and generalization of the m odel. Toward this end, several approaches \nand algorithms have been proposed in the literature , including adversarial training for large \nneural language models (ALUM) [56], generative adver sarial training [57], attacking to training \n(A2T) [58], and large-margin classification [59]. B ias detection and mitigation are important steps \nin ensuring that language models are used ethically and fairly. Here are some approaches to bias \ndetection and mitigation: \n \u2022 Data collection: Bias can be introduced in th e training data that is used to train \nChatGPTs. One approach to reducing bias is to ensur e that the training data is diverse and \nrepresentative of the population that the model wil l be used on. This can involve careful \nselection of data sources and cleaning and preproce ssing the data to remove any biases. \n \u2022 Bias metrics: Once the model is trained, it i s important to measure the extent of any \nbias that may be present. This can be done using va rious bias metrics, such as the disparate \nimpact or statistical parity difference. These metr ics can help identify areas of the model that \nmay be more prone to bias. \n \u2022 Mitigation strategies: Once bias has been ide ntified, there are various strategies \nthat can be used to mitigate it. One approach is to modify the training data to reduce bias, for \nexample by oversampling underrepresented groups. An other approach is to modify the model \nitself, for example by adding constraints or penalt ies to the training process that encourage \nfairer predictions. Alternatively, post-processing techniques can be used to adjust the model\u2019s \npredictions to ensure fairness. \n \u2022 Regular monitoring: Bias detection and mitiga tion is an ongoing process, and it is \nimportant to regularly monitor the model for any ne w sources of bias that may emerge. This \ncan involve setting up automated monitoring systems that flag potential bias in real time, as \nwell as regular audits of the model\u2019s performance. \n \n \n4.3 Multimodal Integration \n Researchers can explore the integration of ChatGPT s with other modalities such as \nimages and videos to enhance their applications in fields such as education and healthcare. \nMultimodal integration is an important research dir ection for language models, as it involves \ncombining multiple modalities of information, such as text, images, and audio, to generate more \ncomprehensive and accurate outputs. Multimodal inte gration can help Chatbots better \nunderstand and respond to complex inputs, and can e nable more natural and intuitive \ninteractions between humans and machines. \nOne approach to multimodal integration in ChatGPTs is", "doc_id": "6012fe95-dfd4-477c-b0c5-98be7f19251a", "embedding": null, "doc_hash": "01eab58188ce2f2e31bbecdc4f945f59e1910edf9e2ec4f284bba450965cb5f9", "extra_info": null, "node_info": {"start": 51245, "end": 55273}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "0780cde1-9a19-4cbd-9525-12a03b247a78", "3": "9aa77e4d-8cdc-4d30-98f1-053ec73fcb4a"}}, "__type__": "1"}, "9aa77e4d-8cdc-4d30-98f1-053ec73fcb4a": {"__data__": {"text": "automated monitoring systems that flag potential bias in real time, as \nwell as regular audits of the model\u2019s performance. \n \n \n4.3 Multimodal Integration \n Researchers can explore the integration of ChatGPT s with other modalities such as \nimages and videos to enhance their applications in fields such as education and healthcare. \nMultimodal integration is an important research dir ection for language models, as it involves \ncombining multiple modalities of information, such as text, images, and audio, to generate more \ncomprehensive and accurate outputs. Multimodal inte gration can help Chatbots better \nunderstand and respond to complex inputs, and can e nable more natural and intuitive \ninteractions between humans and machines. \nOne approach to multimodal integration in ChatGPTs is to use a multimodal transformer \narchitecture [60], which incorporates multiple moda lities of input into a single transformer \nmodel. This approach has been used in a number of a pplications, such as image captioning (e.g., \nXGPT [61]) and video question answering (e.g., AVQA [62), with promising results. Another \napproach is to use multimodal fusion techniques to combine the outputs of separate models \ntrained on different modalities [63]. For example, Zhu et al. used separate models for text and \nimages, and then combined their outputs using a fus ion approach based on self-attention [64]. \nIn addition to these approaches, there is ongoing r esearch into developing new methods for \nmultimodal integration in language models. For exam ple, recent work has explored the use of \nreinforcement learning to learn how to weight diffe rent modalities of input based on their \nrelative importance [65]. Another area of research is developing methods to incorporate \nmultimodal inputs that are not synchronous, such as when text and audio inputs are recorded \nseparately [66]. In general, multimodal integration is an important research direction for \nlanguage models, as it enables more flexible and po werful interactions with users and can \nimprove the accuracy and comprehensiveness of the m odel\u2019s output. \n4.4 Transfer Learning \n \nTransfer learning is a research direction that invo lves using pre-trained models to perform \nspecific tasks in various fields. Researchers can e xplore how ChatGPTs can be fine-tuned for \nspecific applications in education, software engine ering, healthcare, and marketing. \nTransfer learning is a powerful technique that has been successfully applied to language \nmodels [66]. Transfer learning refers to the proces s of training a model on one task or domain, \nand then transferring that knowledge to a new task or domain [67]. In the case of Chatbots, \ntransfer learning involves pre-training a model on a large corpus of text data, and then fine-tuning \nthe model on a specific task or domain. Such method has several benefits for ChatGPTs. First, it \ncan help address the issue of limited training data for specific tasks or domains, by allowing the \nmodel to leverage the knowledge it has gained from pre-training on large amounts of data. \nSecond, transfer learning can help reduce the compu tational cost of training a new model from \nscratch, as the pre-trained model can serve as a st arting point for fine-tuning. \nThere are several approaches to transfer learning i n ChatGPTs. One common approach is \nto use a pre-trained model such as GPT-2 or GPT-3, which have been trained on large amounts \nof diverse text data. The pre-trained model can the n be fine-tuned on a specific task or domain, \nsuch as sentiment analysis or question answering, b y further training the model on a smaller \ndataset specific to that task. Another approach is to use transfer learning to adapt a pre-trained \nmodel to a new language. This involves pre-training the model on a large corpus of text data in \nthe new language, and then fine-tuning the model on specific tasks or domains in that language. \nOverall, transfer learning is a powerful technique for language models, as it enables the model \nto leverage knowledge", "doc_id": "9aa77e4d-8cdc-4d30-98f1-053ec73fcb4a", "embedding": null, "doc_hash": "2f2c6bbdd753e9b91656d525c47c7ce63470678033397425231436b5d18ab5cd", "extra_info": null, "node_info": {"start": 55289, "end": 59354}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "6012fe95-dfd4-477c-b0c5-98be7f19251a", "3": "f7b61543-9a3f-4a9e-908a-ebf80a05d53e"}}, "__type__": "1"}, "f7b61543-9a3f-4a9e-908a-ebf80a05d53e": {"__data__": {"text": " \nThere are several approaches to transfer learning i n ChatGPTs. One common approach is \nto use a pre-trained model such as GPT-2 or GPT-3, which have been trained on large amounts \nof diverse text data. The pre-trained model can the n be fine-tuned on a specific task or domain, \nsuch as sentiment analysis or question answering, b y further training the model on a smaller \ndataset specific to that task. Another approach is to use transfer learning to adapt a pre-trained \nmodel to a new language. This involves pre-training the model on a large corpus of text data in \nthe new language, and then fine-tuning the model on specific tasks or domains in that language. \nOverall, transfer learning is a powerful technique for language models, as it enables the model \nto leverage knowledge gained from pre-training on l arge amounts of data, and can help address \nthe issue of limited training data for specific tas ks or domains. \n \n5 Conclusions \n \nChatGPT and other advanced language models/chatbots are powerful disruptive tools \nthat have the potential to revolutionize various fi elds such as education, software engineering, \nhealthcare, and marketing. However, its drawbacks, such as plagiarism, bias and lack of \ntransparency, need to be addressed, and researchers need to explore research directions such \nas explainability, bias detection and mitigation, m ultimodal integration, and transfer learning to \nensure ChatGPTs are used responsibly and thoughtful ly. The work in this paper surveyed the \npossible avenues where language models can positive ly or negatively contribute to that area, \nwhat possible changes need to be made to counter th e negatives or misuse scenarios, and the \nfuture research directions necessary to achieve wid e, effective, and proper deployment. \n \n \n \n \nReferences \n \n \n1. Kung, T.H.; Cheatham, M.; Medenilla, A.; Sillos, C.; Leon, L.D.; Elepa\u00f1o, C.; Madriaga, M.; \nAggabao, R.; Diaz-Candido, G.; Maningo, J.; et al. Performance of ChatGPT on USMLE: \nPotential for AI-Assisted Medical Education Using La rge Language Models 2022 . \nhttps://doi.org/10.1101/2022.12.19.22283643. \n \n2. Weizenbaum, J. Computer power and human reason ; W.H. Freeman: New York, NY, \nUSA, 1976. \n \n3. Wallace, R.S. The Anatomy of A.L.I.C.E. In Parsing the Turing Test ; Springer \nNetherlands, 2007; pp. 181\u2013210. https://doi.org/10 .1007/978-1-4020-6710-5_13. \n \n4. Carpenter, R. Cleverbot \u2014 cleverbot.com. https:// www.cleverbot.com/, 2008. \n[Accessed 27-Feb-2023]. \n \n5. OpenAI. ChatGPT: Optimizing Language Models for Dia logue \u2014 openai.com. \nhttps://openai.com/blog/chatgpt/, 2022. [Accessed 27-Feb-2023]. \n \n6. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhu tdinov, R.R.; Le, Q.V. XLNet: Generalized \nAutoregressive Pretraining for Language Understandin g. In Proceedings of the Advances \nin Neural Information Processing Systems; Wallach, H.; Larochelle, H.; Beygelzimer, A.; \nd\u00c1lch\u00e9 Buc, F.; Fox, E.; Garnett, R., Eds. Curran A ssociates, Inc., 2019, Vol. 32. \n \n7. Lepikhin, D.; Lee, H.; Xu, Y.; Chen, D.; Firat, O.; H uang, Y.; Krikun, M.; Shazeer, N.; Chen, \nZ. GShard: Scaling Giant Models with Conditional C", "doc_id": "f7b61543-9a3f-4a9e-908a-ebf80a05d53e", "embedding": null, "doc_hash": "330b041a424b4099694487e34da27b691ff0ad79c73dc09daa2dac9decbcb70b", "extra_info": null, "node_info": {"start": 59377, "end": 62559}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "9aa77e4d-8cdc-4d30-98f1-053ec73fcb4a", "3": "c5242644-4e19-4413-9b0f-20b838b08201"}}, "__type__": "1"}, "c5242644-4e19-4413-9b0f-20b838b08201": {"__data__": {"text": "Z.; Yang, Y.; Carbonell, J.; Salakhu tdinov, R.R.; Le, Q.V. XLNet: Generalized \nAutoregressive Pretraining for Language Understandin g. In Proceedings of the Advances \nin Neural Information Processing Systems; Wallach, H.; Larochelle, H.; Beygelzimer, A.; \nd\u00c1lch\u00e9 Buc, F.; Fox, E.; Garnett, R., Eds. Curran A ssociates, Inc., 2019, Vol. 32. \n \n7. Lepikhin, D.; Lee, H.; Xu, Y.; Chen, D.; Firat, O.; H uang, Y.; Krikun, M.; Shazeer, N.; Chen, \nZ. GShard: Scaling Giant Models with Conditional C omputation and Automatic Sharding. \nCoRR 2020 , abs/2006.16668 , http://xxx.lanl.gov/abs/2006.16668 [2006.16668]. \n \n8. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen , D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; \nStoyanov, V. RoBERTa: A Robustly Optimized BERT Pr etraining Approach. CoRR 2019 , \nabs/1907.11692 , http://xxx.lanl.gov/abs/1907.11692 [1907.11692]. \n \n9. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang , S.; Matena, M.; Zhou, Y.; Li, W.; Liu, \nP.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. \nJournal of Machine Learning Research 2020 , 21 , 1\u201367. \n \n10. Tapalova, O.; Zhiyenbayeva, N. Artificial Intellig ence in Education: AIEd for Personalised \nLearning Pathways. Electronic Journal of e-Learning 2022 , 20 , 639\u2013653. \nhttps://doi.org/10.34190/ejel.20.5.2597. \n \n11. Kumar, V.; Boulanger, D. Explainable Automated Essa y Scoring: Deep Learning Really \nHas Pedagogical Value. Frontiers in Education 2020 , 5. \nhttps://doi.org/10.3389/feduc.2020.572367. \n \n12. Raychev, V.; Vechev, M.; Yahav, E. Code completion with statistical language models. In \nProceedings of the Proceedings of the 35th ACM SIGP LAN Conference on Programming \nLanguage Design and Implementation. ACM, 2014. \nhttps://doi.org/10.1145/2594291.2594321. \n \n13. Le, K.T.; Rashidi, G.; Andrzejak, A. A methodology f or refined evaluation of neural code \ncompletion approaches. Data Mining and Knowledge Discovery 2022 , 37 , 167\u2013204. \nhttps://doi.org/10.1007/s10618-022-00866-9. \n \n14. Tu, Z.; Su, Z.; Devanbu, P. On the localness of so ftware. In Proceedings of the \nProceedings of the 22nd ACM SIGSOFT International S ymposium on Foundations of \nSoftware Engineering. ACM, 2014. https://doi.org/1 0.1145/2635868.2635875. \n \n15. Allamanis, M.; Brockschmidt, M.; Khademi, M. Learnin g to Represent Programs with \nGraphs, 2017. https://doi.org/10.48550/ARXIV.1711. 00740. \n \n16. Allamanis, M.; Barr, E.T.; Devanbu, P.; Sutton, C. A Survey of Machine Learning for Big \nCode and Naturalness. ACM Computing Surveys 2018 , 51 , 1\u201337. ", "doc_id": "c5242644-4e19-4413-9b0f-20b838b08201", "embedding": null, "doc_hash": "c6275dfe81b828294b7f07fe6471694c37fe5a9ec51b67079f593bc9f71db73d", "extra_info": null, "node_info": {"start": 62799, "end": 65431}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "f7b61543-9a3f-4a9e-908a-ebf80a05d53e", "3": "fb85b6f9-c436-44bb-a256-53b372536df4"}}, "__type__": "1"}, "fb85b6f9-c436-44bb-a256-53b372536df4": {"__data__": {"text": "P. On the localness of so ftware. In Proceedings of the \nProceedings of the 22nd ACM SIGSOFT International S ymposium on Foundations of \nSoftware Engineering. ACM, 2014. https://doi.org/1 0.1145/2635868.2635875. \n \n15. Allamanis, M.; Brockschmidt, M.; Khademi, M. Learnin g to Represent Programs with \nGraphs, 2017. https://doi.org/10.48550/ARXIV.1711. 00740. \n \n16. Allamanis, M.; Barr, E.T.; Devanbu, P.; Sutton, C. A Survey of Machine Learning for Big \nCode and Naturalness. ACM Computing Surveys 2018 , 51 , 1\u201337. \nhttps://doi.org/10.1145/3212695. \n \n17. Adlung, L.; Cohen, Y.; Mor, U.; Elinav, E. Machine learning in clinical decision making. \nMed 2021 , 2, 642\u2013665. https://doi.org/10.1016/j.medj.2021.04. 006. \n18. Wu, L.; Chen, S.; Guo, L.; Shpyleva, S.; Harris, K.; F ahmi, T.; Flanigan, T.; Tong, W.; Xu, J.; \nRen, Z. Development of benchmark datasets for text mining and sentiment analysis to \naccelerate regulatory literature review. Regulatory Toxicology and Pharmacology \n2023 , 137 , 105287. https://doi.org/10.1016/j.yrtph.2022.105 287. \n \n19. Lederman, A.; Lederman, R.; Verspoor, K. Tasks as nee ds: reframing the paradigm of \nclinical natural language processing research for r eal-world decision support. Journal of \nthe American Medical Informatics Association 2022 , 29 , 1810\u20131817. \nhttps://doi.org/10.1093/jamia/ocac121. \n \n20. Gururangan, S.; Marasovic, A.; Swayamdipta, S.; Lo, K.; Beltagy, I.; Downey, D.; Smith, \nN.A. Don\u2019t Stop Pretraining: Adapt Language Models to Domains and Tasks. CoRR \n2020 , abs/2004.10964 , http://xxx.lanl.gov/abs/2004.10964 [2004.10964]. \n \n21. Liu, Z.; Roberts, R.A.; Lal-Nag, M.; Chen, X.; Huang, R.; Tong, W. AI-based language \nmodels powering drug discovery and development. Drug Discovery Today 2021 , 26 , \n2593\u20132607. https://doi.org/10.1016/j.drudis.2021.0 6.009. \n \n22. Tripathi, A.; Misra, K.; Dhanuka, R.; Singh, J.P. A rtificial Intelligence in Accelerating Drug \nDiscovery and Development. Recent Patents on Biotechnology 2023 , 17 , 9\u201323. \nhttps://doi.org/10.2174/1872208316666220802151129. \n \n23. Bhatnagar, R.; Sardar, S.; Beheshti, M.; Podichetty , J.T. How can natural language \nprocessing help model informed drug development?: a review. JAMIA Open 2022 , 5. \nhttps://doi.org/10.1093/jamiaopen/ooac043. \n \n24. Leyens, L.; Reumann, M.; Malats, N.; Brand, A. Use o f big data for drug development \nand for public and personal health and care. Genetic Epidemiology 2016 , 41 , 51\u201360. \nhttps://doi.org/10.1002/gepi.22012. \n \n25. Ball, R.; Pan, G.D. \u201cArtificial Intelligence\u201d for", "doc_id": "fb85b6f9-c436-44bb-a256-53b372536df4", "embedding": null, "doc_hash": "a7c9c327eaa28c1a37685a7405789b3a4fcd275c0c77b8a15637fb7d3994b06f", "extra_info": null, "node_info": {"start": 65406, "end": 68047}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "c5242644-4e19-4413-9b0f-20b838b08201", "3": "25027da7-f677-499d-90e0-5665fa5133c5"}}, "__type__": "1"}, "25027da7-f677-499d-90e0-5665fa5133c5": {"__data__": {"text": " \n \n23. Bhatnagar, R.; Sardar, S.; Beheshti, M.; Podichetty , J.T. How can natural language \nprocessing help model informed drug development?: a review. JAMIA Open 2022 , 5. \nhttps://doi.org/10.1093/jamiaopen/ooac043. \n \n24. Leyens, L.; Reumann, M.; Malats, N.; Brand, A. Use o f big data for drug development \nand for public and personal health and care. Genetic Epidemiology 2016 , 41 , 51\u201360. \nhttps://doi.org/10.1002/gepi.22012. \n \n25. Ball, R.; Pan, G.D. \u201cArtificial Intelligence\u201d for Pharmacovigilance: Ready for Prime Time? \nDrug Safety 2022 , 45 , 429\u2013438. https://doi.org/10.1007/s40264-022-0115 7-4. \n \n26. Koneti, G.; Das, S.S.; Bahl, J.; Ranjan, P.; Ramamur thi, N. Discovering the Knowledge in \nUnstructured Early Drug Development Data Using NLP a nd Advanced Analytics. In \nProceedings of the 2022 IEEE International Conferen ce on Bioinformatics and \nBiomedicine (BIBM), 2022, pp. 3840\u20133842. \nhttps://doi.org/10.1109/BIBM55620.2022.9995435. \n \n27. Wang, X.; Xu, X.; Tong, W.; Liu, Q.; Liu, Z. DeepCa usality: A general AI-powered causal \ninference framework for free text: A case study of LiverTox. Frontiers in Artificial \nIntelligence 2022 , 5. https://doi.org/10.3389/frai.2022.999289. \n \n28. Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C.H .; Kang, J. BioBERT: a pre-trained \nbiomedical language representation model for biomed ical text mining. Bioinformatics \n2019 , 36 , 1234\u20131240. https://doi.org/10.1093/bioinformatic s/btz682. \n \n29. Huang, K.; Altosaar, J.; Ranganath, R. ClinicalBERT : Modeling Clinical Notes and \nPredicting Hospital Readmission. CoRR 2019 , abs/1904.05342 , \nhttp://xxx.lanl.gov/abs/1904.05342 [1904.05342]. \n \n30. Alsentzer, E.; Murphy, J.; Boag, W.; Weng, W.H.; Ji ndi, D.; Naumann, T.; McDermott, M. \nPublicly Available Clinical. In Proceedings of the Proceedings of the 2nd Clinical Natural \nLanguage Processing Workshop. Association for Comput ational Linguistics, 2019. \nhttps://doi.org/10.18653/v1/w19-1909. \n \n31. Verma, S.; Sharma, R.; Deb, S.; Maitra, D. Artific ial intelligence in marketing: Systematic \nreview and future research direction. International Journal of Information \nManagement Data Insights 2021 , 1, 100002. \nhttps://doi.org/10.1016/j.jjimei.2020.100002. \n \n32. Nguyen, Q.N.; Sidorova, A.; Torres, R. User intera ctions with chatbot interfaces vs. \nMenu-based interfaces: An empirical study. Computers in Human Behavior 2022 , 128 , \n107093. https://doi.org/10.1016/j.chb.2021.107093. \n \n33. Chipman, S. What is CRM artificial intelligence an d what can it do for my business? - \nCRM Switch \u2014 crmswitch.com, 2023. [Accessed 06-Mar -2023]. \n \n34. ", "doc_id": "25027da7-f677-499d-90e0-5665fa5133c5", "embedding": null, "doc_hash": "f53f49e06053ca0c480ae781ff6177a173e3f22217f1969f62f1bc5ba098db3d", "extra_info": null, "node_info": {"start": 68074, "end": 70786}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "fb85b6f9-c436-44bb-a256-53b372536df4", "3": "e62b0869-d6f1-4863-9be0-173f9fbfae9d"}}, "__type__": "1"}, "e62b0869-d6f1-4863-9be0-173f9fbfae9d": {"__data__": {"text": " International Journal of Information \nManagement Data Insights 2021 , 1, 100002. \nhttps://doi.org/10.1016/j.jjimei.2020.100002. \n \n32. Nguyen, Q.N.; Sidorova, A.; Torres, R. User intera ctions with chatbot interfaces vs. \nMenu-based interfaces: An empirical study. Computers in Human Behavior 2022 , 128 , \n107093. https://doi.org/10.1016/j.chb.2021.107093. \n \n33. Chipman, S. What is CRM artificial intelligence an d what can it do for my business? - \nCRM Switch \u2014 crmswitch.com, 2023. [Accessed 06-Mar -2023]. \n \n34. Mauro, A.D.; Sestino, A.; Bacconi, A. Machine lear ning and artificial intelligence use in \nmarketing: a general taxonomy. Italian Journal of Marketing 2022 , 2022 , 439\u2013457. \nhttps://doi.org/10.1007/s43039-022-00057-w. \n \n35. Duarte, V.; Zuniga-Jara, S.; Contreras, S. Machine Learning and Marketing: A Systematic \nLiterature Review. IEEE Access 2022 , 10 , 93273\u201393288. \nhttps://doi.org/10.1109/access.2022.3202896. \n \n36. Bruyn, A.D.; Viswanathan, V.; Beh, Y.S.; Brock, J.K. U.; von Wangenheim, F. Artificial \nIntelligence and Marketing: Pitfalls and Opportunit ies. Journal of Interactive Marketing \n2020 , 51 , 91\u2013105. https://doi.org/10.1016/j.intmar.2020.04 .007. \n \n37. Schramowski, P.; Turan, C.; Andersen, N.; Rothkopf, C.A.; Kersting, K. Large pre-trained \nlanguage models contain human-like biases of what i s right and wrong to do. Nature \nMachine Intelligence 2022 , 4, 258\u2013268. https://doi.org/10.1038/s42256-022-0045 8-8. \n \n \n38. Rudin, C. Stop explaining black box machine learni ng models for high stakes decisions \nand use interpretable models instead. Nature Machine Intelligence 2019 , 1, 206\u2013215. \nhttps://doi.org/10.1038/s42256-019-0048-x. \n \n \n39. Yang, X. An Overview of the Attention Mechanisms i n Computer Vision. Journal of \nPhysics: Conference Series 2020 , 1693 , 012173. https://doi.org/10.1088/1742-\n6596/1693/1/012173. \n \n40. An, J.; Joe, I. Attention Map-Guided Visual Explan ations for Deep Neural Networks. \nApplied Sciences 2022 , 12 , 3846. https://doi.org/10.3390/app12083846. \n \n41. Hicks, S.A.; Isaksen, J.L.; Thambawita, V.; Ghouse, J.; Ahlberg, G.; Linneberg, A.; Grarup, \nN.; Str\u00fcmke, I.; Ellervik, C.; Olesen, M.S.; et al. Explaining deep neural networks for \nknowledge discovery in electrocardiogram analysis. Scientific Reports 2021 , 11 . \nhttps://doi.org/10.1038/s41598-021-90285-5. \n \n42. Riegler, M.; Lux, M.; Griwodz, C.; Spampinato, C.; d e Lange, T.; Eskeland, S.L.; Pogorelov, \nK.; Tavanapong, W.; Schmidt, P.T.; Gurrin, C.; et al . Multimedia and Medicine. In", "doc_id": "e62b0869-d6f1-4863-9be0-173f9fbfae9d", "embedding": null, "doc_hash": "688488b63effcb96f645323bed3441480a9d9ac0aec1b9cad761890e6a14c33c", "extra_info": null, "node_info": {"start": 70756, "end": 73390}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "25027da7-f677-499d-90e0-5665fa5133c5", "3": "f2a49d34-de3c-4aa1-8081-4ef7b9e79a60"}}, "__type__": "1"}, "f2a49d34-de3c-4aa1-8081-4ef7b9e79a60": {"__data__": {"text": "V.; Ghouse, J.; Ahlberg, G.; Linneberg, A.; Grarup, \nN.; Str\u00fcmke, I.; Ellervik, C.; Olesen, M.S.; et al. Explaining deep neural networks for \nknowledge discovery in electrocardiogram analysis. Scientific Reports 2021 , 11 . \nhttps://doi.org/10.1038/s41598-021-90285-5. \n \n42. Riegler, M.; Lux, M.; Griwodz, C.; Spampinato, C.; d e Lange, T.; Eskeland, S.L.; Pogorelov, \nK.; Tavanapong, W.; Schmidt, P.T.; Gurrin, C.; et al . Multimedia and Medicine. In \nProceedings of the Proceedings of the 24th ACM inte rnational conference on \nMultimedia. ACM, 2016. https://doi.org/10.1145/296 4284.2976760. \n \n43. Kelly, C.J.; Karthikesalingam, A.; Suleyman, M.; Corr ado, G.; King, D. Key challenges for \ndelivering clinical impact with artificial intellig ence. BMC Medicine 2019 , 17 . \nhttps://doi.org/10.1186/s12916-019-1426-2. \n \n44. Chen, D.; Liu, S.; Kingsbury, P.; Sohn, S.; Storlie, C.B.; Habermann, E.B.; Naessens, J.M.; \nLarson, D.W.; Liu, H. Deep learning and alternative learning strategies for retrospective \nreal-world clinical data. npj Digital Medicine 2019 , 2. https://doi.org/10.1038/s41746-\n019-0122-0. \n \n45. Khasawneh, N.; Fraiwan, M.; Fraiwan, L. Detection of K-complexes in EEG waveform \nimages using faster R-CNN and deep transfer learnin g. BMC Medical Informatics and \nDecision Making 2022 , 22 . https://doi.org/10.1186/s12911-022-02042-x. \n \n \n46. Khasawneh, N.; Fraiwan, M.; Fraiwan, L. Detection of K-complexes in EEG signals using \ndeep transfer learning and YOLOv3. Cluster Computing 2022 . \nhttps://doi.org/10.1007/s10586-022-03802-0. \n \n \n47. Samek, W.; Montavon, G.; Vedaldi, A.; Hansen, L.K.; M\u00fcller, K.R. Explainable AI: \nInterpreting, explaining and visualizing deep learn ing ; Springer Nature: Cham, \nSwitzerland, 2019. \n \n \n48. Samek, W.; M\u00fcller, K.R. Towards Explainable Artific ial Intelligence. In Explainable AI: \nInterpreting, Explaining and Visualizing Deep Learni ng ; Springer International Publishing, \n2019; pp. 5\u201322. https://doi.org/10.1007/978-3-030- 28954-6_1. \n \n \n49. Singh, A.; Sengupta, S.; Lakshminarayanan, V. Expla inable Deep Learning Models in \nMedical Image Analysis. Journal of Imaging 2020 , 6, 52. \nhttps://doi.org/10.3390/jimaging6060052. \n \n \n50. Bai, X.; Wang, X.; Liu, X.; Liu, Q.; Song, J.; Sebe, N.; Kim, B. Explainable deep learning for \nefficient and robust pattern recognition: A survey of recent developments. Pattern \nRecognition 2021 , 120 , 108102. https://doi.org/10.1016/j.patcog.2021.10 8102. \n \n \n51. Choo, J.; Liu, S. Visual Analytics for Explainable Deep Learning. ", "doc_id": "f2a49d34-de3c-4aa1-8081-4ef7b9e79a60", "embedding": null, "doc_hash": "eb2e1fe28f046e882d541656cb9923f4adb8044a85c857d88950e044c9e41f77", "extra_info": null, "node_info": {"start": 73451, "end": 76075}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "e62b0869-d6f1-4863-9be0-173f9fbfae9d", "3": "bf73a3f1-3f4f-4c5d-8809-5acbede2d600"}}, "__type__": "1"}, "bf73a3f1-3f4f-4c5d-8809-5acbede2d600": {"__data__": {"text": "S.; Lakshminarayanan, V. Expla inable Deep Learning Models in \nMedical Image Analysis. Journal of Imaging 2020 , 6, 52. \nhttps://doi.org/10.3390/jimaging6060052. \n \n \n50. Bai, X.; Wang, X.; Liu, X.; Liu, Q.; Song, J.; Sebe, N.; Kim, B. Explainable deep learning for \nefficient and robust pattern recognition: A survey of recent developments. Pattern \nRecognition 2021 , 120 , 108102. https://doi.org/10.1016/j.patcog.2021.10 8102. \n \n \n51. Choo, J.; Liu, S. Visual Analytics for Explainable Deep Learning. IEEE Computer Graphics \nand Applications 2018 , 38 , 84\u201392. https://doi.org/10.1109/MCG.2018.04273166 1. \n \n \n52. Liu, Y.; Li, H.; Guo, Y.; Kong, C.; Li, J.; Wang, S. Rethinking Attention-Model Explainability \nthrough Faithfulness Violation Test. CoRR 2022 , abs/2201.12114 , \nhttp://xxx.lanl.gov/abs/2201.12114 [2201.12114]. \n \n \n53. Ribeiro, M.T.; Singh, S.; Guestrin, C. \"Why Should I Trust You?\": Explaining the \nPredictions of Any Classifier. CoRR 2016 , abs/1602.04938 , \nhttp://xxx.lanl.gov/abs/1602.04938 [1602.04938]. \n \n \n54. Aditya, P.S.R.; Pal, M. Local Interpretable Model Agnostic Shap Explanations for \nmachine learning models, 2022. https://doi.org/10. 48550/ARXIV.2210.04533. \n \n \n55. Zhao, W.; Alwidian, S.; Mahmoud, Q.H. Adversarial Training Methods for Deep Learning: \nA Systematic Review. Algorithms 2022 , 15 , 283. https://doi.org/10.3390/a15080283. \n \n56. Liu, X.; Cheng, H.; He, P.; Chen, W.; Wang, Y.; Poon , H.; Gao, J. Adversarial Training for \nLarge Neural Language Models. CoRR 2020 , abs/2004.08994 , \nhttp://xxx.lanl.gov/abs/2004.08994 [2004.08994]. \n \n \n57. Movahedi, S.; Shakery, A. Generative Adversarial T raining Can Improve Neural Language \nModels, 2022. https://doi.org/10.48550/ARXIV.2211. 09728. \n \n \n58. Yoo, J.Y.; Qi, Y. Towards Improving Adversarial Tr aining of NLP Models. In Proceedings \nof the Findings of the Association for Computationa l Linguistics: EMNLP 2021; \nAssociation for Computational Linguistics: Punta Ca na, Dominican Republic, 2021; pp. \n945\u2013956. https://doi.org/10.18653/v1/2021.findings -emnlp.81. \n \n \n59. Wang, D.; Gong, C.; Liu, Q. Improving Neural Languag e Modeling via Adversarial \nTraining. In Proceedings of the Proceedings of the 36th International Conference on \nMachine Learning; Chaudhuri, K.; Salakhutdinov, R., E ds. PMLR, 2019, Vol. 97, \nProceedings of Machine Learning Research , pp. 6555\u20136565. \n \n \n60. Xu, P.; Zhu, X.; Clifton, D.A. Multimodal Learning with Transformers: A Survey, 2022. \nhttps://doi.org/10.48550/ARXIV.2206.06488. \n \n \n61. ", "doc_id": "bf73a3f1-3f4f-4c5d-8809-5acbede2d600", "embedding": null, "doc_hash": "292cdeb011353a6ad2ed315a65f8556f35cc6cbdb8d3497519257890f429292e", "extra_info": null, "node_info": {"start": 76031, "end": 78645}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "f2a49d34-de3c-4aa1-8081-4ef7b9e79a60", "3": "ee3a3718-eeac-4615-808c-897a216b27a5"}}, "__type__": "1"}, "ee3a3718-eeac-4615-808c-897a216b27a5": {"__data__": {"text": "-emnlp.81. \n \n \n59. Wang, D.; Gong, C.; Liu, Q. Improving Neural Languag e Modeling via Adversarial \nTraining. In Proceedings of the Proceedings of the 36th International Conference on \nMachine Learning; Chaudhuri, K.; Salakhutdinov, R., E ds. PMLR, 2019, Vol. 97, \nProceedings of Machine Learning Research , pp. 6555\u20136565. \n \n \n60. Xu, P.; Zhu, X.; Clifton, D.A. Multimodal Learning with Transformers: A Survey, 2022. \nhttps://doi.org/10.48550/ARXIV.2206.06488. \n \n \n61. Xia, Q.; Huang, H.; Duan, N.; Zhang, D.; Ji, L.; Sui , Z.; Cui, E.; Bharti, T.; Zhou, M. XGPT: \nCross-modal Generative Pre-Training for Image Capti oning. In Natural Language \nProcessing and Chinese Computing ; Springer International Publishing, 2021; pp. 786\u2013\n797. https://doi.org/10.1007/978-3-030-88480-2_63. \n \n \n62. Yun, H.; Yu, Y.; Yang, W.; Lee, K.; Kim, G. Pano-AVQA : Grounded Audio-Visual Question \nAnswering on 360 Videos. In Proceedings of the 202 1 IEEE/CVF International \nConference on Computer Vision (ICCV). IEEE, 2021. \nhttps://doi.org/10.1109/iccv48922.2021.00204. \n \n \n63. Paw \u0142owski, M.; Wr\u00f3blewska, A.; Sysko-Roma\u0144czuk, S. Eff ective Techniques for \nMultimodal Data Fusion: A Comparative Analysis. Sensors 2023 , 23 , 2381. \nhttps://doi.org/10.3390/s23052381. \n \n \n64. ] Zhu, H.; Wang, Z.; Shi, Y.; Hua, Y.; Xu, G.; Den g, L. Multimodal Fusion Method Based \non Self-Attention Mechanism. Wireless Communications and Mobile Computing 2020 , \n2020 , 1\u20138. https://doi.org/10.1155/2020/8843186. \n \n65. Bernardino, G.; Jonsson, A.; Loncaric, F.; Castello te, P.M.M.; Sitges, M.; Clarysse, P.; \nDuchateau, N. Reinforcement Learning for Active Mod ality Selection During Diagnosis. \nIn Lecture Notes in Computer Science ; Springer Nature Switzerland, 2022; pp. 592\u2013601. \nhttps://doi.org/10.1007/978-3-031-16431-6_56. \n \n \n66. Bayoudh, K.; Knani, R.; Hamdaoui, F.; Mtibaa, A. A s urvey on deep multimodal learning \nfor computer vision: advances, trends, applications , and datasets. The Visual Computer \n2021 , 38 , 2939\u20132970. https://doi.org/10.1007/s00371-021-02 166-7. \n \n \n67. Fraiwan, M.; Audat, Z.; Fraiwan, L.; Manasreh, T. U sing deep transfer learning to detect \nscoliosis and spondylolisthesis from x-ray images. PLOS ONE 2022 , 17 , e0267851. \nhttps://doi.org/10.1371/journal.pone.0267851. \n \n \n \n ", "doc_id": "ee3a3718-eeac-4615-808c-897a216b27a5", "embedding": null, "doc_hash": "e40fcbd51a88772e814351ed641ac753eff39b3b1d34e6de77941100423b5660", "extra_info": null, "node_info": {"start": 78595, "end": 80951}, "relationships": {"1": "878bdaad-e099-44e7-8cfb-d214c9e407d1", "2": "bf73a3f1-3f4f-4c5d-8809-5acbede2d600"}}, "__type__": "1"}}} |