Introduction

Generative AI models have a long history in artificial intelligence, dating back to the 1950s with the development of Hidden Markov Models (HMMs) and Gaussian Mixture Models (GMMs: Cao et al., 2023). However, the November 2022 release of ChatGPT by Open AI has accelerated widespread global awareness and the use of generative AI in multiple ways (Metz, 2022; Reed, 2022; Tung, 2023). Higher education faculty are among those who have followed the release of ChatGPT with a great deal of interest (Dwivedi et al., 2023). Strong sentiments are being expressed, both positive and negative, about the use of ChatGPT in colleges and universities (e.g., Chronicle of Higher Education, 2023). Faculty find themselves in uncertain territory regarding the affordances and the challenges that generative AI presents. This uncertainty has caused a range of responses from faculty regarding their perception of ChatGPT. However, at this time no research has been done to determine how faculty perceive the use of ChatGPT. Nor has there been any research to determine the factors that cause those perceptions. In response to this paucity of research, scholars (e.g., Dwivedi et al., 2023; Haque et al., 2022; Leiter et al., 2023; Taecharungroj, 2023; Tlili et al., 2023) are calling for more investigation into the impact of ChatGPT in the education sector. The purpose of this study was to respond to this call and explore higher education faculty perceptions regarding the use of ChatGPT and investigate what factors are causing these perceptions.

Literature Review

ChatGPT is an artificial intelligence chatbot created by Open AI and released in November 2022. As a large advanced natural language processing model pre-trained on a large dataset, ChatGPT can generate natural language that is coherent, contextually appropriate, and human-like. It uses a transformer architecture that merges the benefits of generative models with scalability enabling text production in any terminology and topic from a specified prompt (Aydın & Karaarslan, 2023). Currently, ChatGPT has over 100 million users and the ChatGPT website receives an estimated 1 billion monthly visitors (Milmo, 2023). The world-wide interest in this technology far surpasses any other modern technology in terms of adoption rates (Reuters, 2023). While relatively new, ChatGPT’s affordances and challenges have been discussed, analyzed, praised, and disparaged by popular media, experts, and laymen alike, such as the Wall Street Journal (Hao, 2023).

Affordances of ChatGPT in Higher Education

Proponents of ChatGPT in higher education state that ChatGPT is a technology with rich potential to empower students and enhance their educational experiences and resources (Dai et al., 2023). Some of the general affordances of ChatGPT are the ability to promote lifelong learning, provide alternative ideas, be a creative source of inspiration, improve human capacity for knowledge access and development, form personalized learning pathways, encourage competence building, and serve as a conversational agent that can be used to simulate social interactions (Bozkurt et al., 2023). Furthermore, ChatGPT can provide affordances in specific content areas. For instance, in language learning, the generation of realistic dialogues can provide learners with authentic language use (George & George, 2023). ChatGPT’ s language translation capacity allows learners to input text in their native language and have it automatically translated into their target language (Jiao et al., 2023). ChatGPT can also generate writing prompts and provide feedback on written work, leading to improvement in learners’ writing skills (Baidoo-Anu & Owusu Ansah, 2023).

Challenges and Limitations of ChatGPT in Higher Education

Although the potential for positive use of ChatGPT in higher education is high, there are concomitant challenges. One challenge is the concern regarding plagiarism and academic integrity. AI writing systems are designed to generate text based on a set of parameters or prompts. Students could potentially use these systems to cheat on their assignments by submitting text created by ChatGPT as their own work (Dehouche, 2021). Furthermore, ChatGPT, which is effectively a text generating machine, does not grasp the information it generates, nor does it judge whether it is accurate or relevant, possibly leading to the creation and distribution of false and misleading information (Mhlanga, 2023).

Large language models (LLMs), such as ChatGPT, can perpetuate and amplify existing biases and unfairness in society (Farrokhnia et al., 2023). Typically, LLMs rely heavily on training data, and when the data contain biases or anomalies, it could lead to unfair or biased results (Kasneci et al., 2023). Internal mechanisms and processes in generative AI models are not sufficiently open and transparent to users about how they work. Therefore, it is important to ensure that the decision-making processes of these models are transparent to users. ChatGPT and other generative models involve the collection and processing of personal data, which raises concerns about privacy and data security (Rahman & Watanobe, 2023). There is concern of the potential to depend excessively on ChatGPT as a tool, which could hinder the development of key professional and academic skills by both students and teachers (Kasneci et al., 2023).

Importance of Faculty Perceptions in Adoption of Technology

Although ChatGPT will impact all facets of life, many academics assert that teaching, learning, and academic research are likely to experience some of the most transformative impacts of ChatGPT (Dwivedi et al., 2023). It is well-documented that educators in the past have not made effective use of instructional technologies (Bingimlas, 2009; Gautreau, 2011; Keengwe & Kang, 2012). This lack of evidence supporting the effective integration of technology by educators has prompted researchers to call for the need to identify factors influencing faculty adoption of instructional technology (Bingimlas, 2009). One factor that is important to understand is how faculty perceive the integration of classroom technology tools (Burch & Mohammed, 2019). Perception is the process by which an individual selects, organizes and interprets stimuli into a meaningful and coherent picture of the world. It is a cognitive process by which people attend to incoming stimuli, organize, and interpret such stimuli into behavior (McDonald, 2012). Understanding how higher education faculty perceive ChatGPT can help to ensure more effective integration of this technology.

Related Work

ChatGPT has only been available since November 2022 providing only a short period time in which academic literature could be published regarding faculty perceptions of the use of ChatGPT. However, a review of the research reveals a few early articles that are relevant to this study.

Taecharungroj (2023) collected and analyzed tweets about ChatGPT in English from November 30 to December 31, 2022, with the purpose of identifying the perceptions of early general users of ChatGPT. The Latent Dirichlet Allocation (LDA) topic modeling algorithm was used to answer the question “What can ChatGPT do?” The results revealed topics that were grouped into three domains: general topics, functional domains, and potential impacts. The general topics included news reporting about ChatGPT, the technology behind it, and global reactions. The functional domains referred to five possible functions of ChatGPT, creative writing, essay writing, prompt writing, code writing, and answering questions. The potential impacts domain included two topics, impacting tech, and impacting humans. Taecharungroj specifically calls for more investigation into the impact of ChatGPT in the education sector to best prepare educators for the rapidly changing landscape of ChatGPT.

Haque et al. (2022), explored the sentiments of ChatGPT early adopters using 10,732 tweets from December 5, 2022, to December 7, 2022. These researchers used topic modelling and sentiment analysis to answer the following questions. “What are the characteristics of ChatGPT early adopters? What are the main topics that were discussed? What are the sentiments that were expressed about ChatGPT topics?” These researchers found that the USA, India, UK, Canada, and Germany were the top-5 countries in which people expressed their opinion of ChatGPT. The top 3 user occupations were software practitioners, academics, and students. Nine topics were identified as being discussed by early adopters. They are disruptions to software development, entertainment and exercising creativity, natural language processing, impact on educational aspects, chatbot intelligence, impact on business development, implications for search engines, Q &A testing, and future careers and opportunities. Results of users’ sentiment regarding ChatGPT were reported by topic as positive, negative, and neutral. This study revealed that the adoption of ChatGPT for educational purposes reported both positive and negative perceptions among the users (52% positive, 32% negative and 16% neutral).

In another study, Leiter et al. (2023) expanded the literature about the perception of the use of ChatGPT by analyzing 330 K X tweets and more than 150 scientific papers from Arxiv and Semantic scholar through February 9, 2023. This data set included tweets from over 61 languages. The study sought to answer these questions. “How is ChatGPT viewed from the perspectives of different actors?” “How have perceptions changed over time, and what limitations and strengths were pointed out?” The results indicated that ChatGPT is generally perceived positively with emotion of joy being the most prevalent. Since its inception, the positive perception has decreased slightly and in languages other than English. In the analysis of data specifically related to education, the number of papers that see ChatGPT as a threat is almost equal to the number of those who view it an opportunity. These authors also call for more research, particularly calling for investigating the perceptions from specific user groups. These three studies all provide an overarching perspective as to how all users perceive ChatGPT.

Providing a more specific look at perceptions regarding ChatGPT used in an educational context, Tlili et al. (2023), focused on identifying the concerns of using ChatGPT. They used a three-stage instrumental case study, with a social network analysis of tweets, content analysis of interviews, and investigation of user experience the authors analyzed 2330 tweets, from 1530 X users from December 23, 2022, to January 6, 2023. In addition, interviews were conducted with 19 users who have been using ChatGPT in education to investigate how different stakeholders (students, educators, etc.) perceive the use of ChatGPT in education, especially their concerns. The study results revealed that the frequency of positive sentiments was almost twice as high as the negative ones. The content analysis of the interviews revealed that the users found ChatGPT very significant with a great value to revolutionize education; however, they raised several concerns calling for caution and the need for more guidelines about how to use ChatGPT safely in education. This study is helpful in providing a focus on the educational context. However, this study involved educators who worked in a variety of contexts.

Sullivan et al. (2023), conducted a study focused on higher education to provide a content analysis of 100 news articles from Australia, New Zealand, the United States and the United Kingdom. This may reflect some of the concerns shared by faculty but are not directly from faculty. The news text was coded to explore several themes in relation to the impact of ChatGPT on higher education, including university responses, academic integrity concerns, the limitations and weaknesses of AI tool outputs, and opportunities for student learning. The findings revealed that the public discussion and university responses about ChatGPT have focused mainly on academic integrity concerns and innovative assessment design. It also revealed a lack of a student voice in the conversation.

These extant studies provide some helpful background as to the perceptions of higher education faculty on ChatGPT, but to date a study has not been conducted to explore faculty perceptions. It is important to note that the authors of all the studies cited above call for more investigation into the impact ChatGPT will have on how teachers teach, and students learn as artificial intelligence becomes more ubiquitous in our world.

Purpose of this Study

The purpose of this study is in response to the appeal from scholars (e.g., Dwivedi et al., 2023; Haque et al., 2022; Leiter et al., 2023; Taecharungroj, 2023; Tlili et al., 2023) to investigate the use of ChatGPT in higher education. The overarching research question for this study is: What are the perceptions of higher education faculty regarding the use of ChatGPT? Three individual sub-questions help to provide a more in-depth and specific understanding of higher education’s faculty’s perception of a technology that has the potential to revolutionize education (Malik et al., 2023).

  1. 1.

    What are the sentiments that are being expressed about ChatGPT by higher education faculty on X?

  2. 2.

    How have these sentiments changed over time?

  3. 3.

    What factors influence a faculty member to perceive ChatGPT in a specific way?

Method

In recent years, there has been a significant increase in researchers utilizing social media platforms, particularly X, to extract insights and gain a deeper understanding of emerging research topics, such as ChatGPT (Fiesler & Proferes, 2018). This study used a purposeful sampling procedure to select higher education faculty members who utilized X to discuss ChatGPT. A sentiment analysis was then conducted to classify faculty opinions about ChatGPT (Hutto & Gilbert, 2014; Liu et al., 2017). Subsequently, this study expanded beyond simple sentiment analysis by integrating emotional analysis, enabling the extraction of a diverse range of emotions from the tweets (Lyu et al., 2021; Mohammad & Turney, 2013). Finally, to extract insights and generate knowledge regarding factors that influence faculty perceptions of ChatGPT, we conducted grounded coding on the categorized emotional sentiments. Figure 1 provides an overview of the research methodology for this study.

Fig. 1
figure 1

Overview of the Research Methodology

Data Collection

The X Application Programming Interface (API) was utilized to gather a comprehensive corpus of historical tweets posted by X accounts from November 30th, 2022, to April 30th, 2023. Following the recommendation of Humphreys and Wang (2018) emphasizing the importance of selecting an appropriate platform and employing suitable sampling techniques for textual analysis, X was chosen as the platform due to its ability to facilitate virtual discussions and create issue-specific virtual town halls (O’Hallarn et al., 2018). Tweets were collected that included the keyword words “#ChatGPT” and “#highered”. The hashtag “#ChatGPT” is commonly used on social media to engage in conversations related to artificial intelligence. There are three reasons for choosing the hashtag #highered. Firstly, it holds significance in academic discourse, particularly evident in discussions related to crisis management during the COVID-19 pandemic among faculties and administrators (Biddix et al., 2023). Secondly, an analysis of the top 10 hashtags in education revealed that #highered accounted for 36% of usage, which is 2.5 times higher than the second most popular hashtag, #highereducation. Thirdly, with access to the full X archive for data collection, we examined the top three hashtags (#highered, #highereducation, and #education) and found that #highered represented the majority of discussions among faculty and administrators (Barrie & Ho, 2021) R code was written to access terms get_all_tweets (query = “#ChatGpt #highered”, lang = en, start_tweets = “2022-11-30T00:00:00Z”, end_tweets = “2023-04-30T00:00:00Z”, granularity = “day”, n = 151). The search resulted in a total of 3559 comments, including original tweets and retweets.

Data Preprocessing

X data underwent preprocessing, manipulation, cleaning, formatting, and filtering to eliminate irrelevant content from the study. First, non-English language tweets were filtered out. Next, duplicate tweets from individual users were eliminated, and manual verification was conducted to identify bots based on account usage, sentiment, and social network characteristics (Rodríguez-Ruiz et al., 2020). Further, noise removal included removing tweets containing URLs and symbols and the removal of stop words such as “the,” “a,” and “to,” which did not help to distinguish one tweet from another. Only distinct tweets from each user were deemed appropriate to be included. Lastly, tweets were annotated with tokenization and parts of speech. After applying data preprocessing techniques, a total of 2115 unique users’ tweets were available for analysis.

Sentiment Analysis

Sentiment analysis is a new field of research born in Natural Language Processing (NLP) used to detect subjectivity in text and/or extract and classify sentiments and emotions (Liu et al., 2017). Sentiment analysis examines individuals’ opinions and emotions concerning events, technology, social issues, and experiences (Mamo, 2023). Sentiment analysis aims to assess the level of positivity or negativity in written content related to a specific subject, such as social media posts, product reviews, news articles, and blogs (Berger et al., 2020; Humphrey & Wang, 2018; Kauffmann et al., 2020). The analysis of sentiment in user-generated content has received considerable attention in academic research across various disciplines such as marketing (Berger et al., 2020), management (Etter et al., 2018), and consumer behavior (Humphrey & Wang, 2018). There are two main approaches to performing sentiment analysis: (1) lexicon-based/dictionary-based and (2) machine learning approaches (Humphrey & Wang, 2018). This study chose to use the lexicon/dictionary-based sentiment analysis. This approach is widely recognized for its effectiveness in detecting sentiments (Shivaprasad & Shetty, 2017; Mamo, 2023). Humphrey and Wang (2018) contend that when the intended construct, such as positive and negative affect, is relatively well-defined, a lexicon/dictionary-based technique is the most suitable analysis method. Additionally, dictionary-based sentiment analysis is validated in various settings and provides transparent and easily interpretable results (Berger et al., 2020).

The Valence Aware Dictionary for Sentiment Reasoning VADER was used to analyze the sentiments in the 2115 tweets in the dataset in this study. VADER is the gold standard for lexicon-based sentiment analysis in short texts like X (Al-Natour & Turetken, 2020; Hutto & Gilbert, 2014; Mamo & Anagnostopoulos, 2023; Riberio et al., 2016) and is used across various types of textual content, including movie reviews, newspaper articles, X posts, and Amazon product reviews (Hutto & Gilbert, 2014).The VADER analysis assigned a polarity score for each of the domains of negative, positive, neutral, or compound. Among these polarity values, the compound score is particularly useful for measuring the sentiment in a document (Elbagir & Yang, 2019). A compound score greater than 0.05 indicated positive sentiment, a compound score less than − 0.05 indicated negative sentiment, and compound scores within the range of -0.05 to 0.05 were considered neutral. In addition, the National Research Council Word-Emotion Association Lexicon, commonly referred to as the NRC Emotion Lexicon or EmoLex was used in this study. The NRC is a lexicon-based dictionary used in text analysis to identify specific emotional content in written social media posts. It provides annotations and associations with eight basic emotions, namely anger, fear, anticipation, trust, surprise, sadness, joy, and disgust. The NRC Emotion Lexicon was utilized to analyze text data and determine the presence or intensity of these emotions within the text (Lyu et al., 2021).

It is important to note that any type of quantitative model of languages is not globally the best method and needs to be validated (Grimmer & Stewart, 2013). Human validation is considered the gold standard, and researchers often manually double-check the validity of their classifications. To determine the threshold for categorizing the scores for this study, the distribution and reliability of the compound sentiment score produced by VADER sentiment dictionary were examined. Following the suggestion of O’Connor and Joffe (2020), a random sample comprising 20% of the data was manually coded by two co-authors who were not involved in the dictionary-based sentiment analysis classification. The coding was then reviewed by another co-author with expertise in big data and social media analytics. Subsequently, all co-authors engaged in face-to-face discussions to cross-check the manually coded sentiment and VADER scores. The outcome of the discussions yielded an agreement percentage of 90%, surpassing the acceptable reliability threshold recommended by Miles and Huberman (1994). Two authors discussed the remaining 10% of the lexicon-based results and resolved any inconsistencies in the value labeling.

Grounded Coding

With the data gathered from the sentiment analysis, grounded coding (Strauss & Corbin, 1995). was then used to identify the factors that influence faculty perceptions. For each faculty sentiment uncovered in the sentiment analysis, the earlier analysis provided a set of corresponding Tweets. The grounded coding design used a constant comparative method to identify important text from the tweets that illustrated the factors that led to specific sentiments. Through an iterative, inductive process, codes were developed, and with a constant comparison of factors with factors, of factors with codes, and codes with codes. Codes were deemed theoretically saturated when all the factors fit with one of the specific sentiments. Two researchers coded and reached an inter-rater percentage agreement of 100% on the factors. “In vivo” (Saldana, 2015) coding was used within the grounded coding. In vivo coding involves using language from the original Tweet to ensure that the faculty members’ intentions were accurately captured.

Results

RQ1: What are the Sentiments that are being Expressed about ChatGPT by Higher Education Faculty on X?

Research question one focuses on the sentiments expressed by higher education faculty on X regarding ChatGPT from November 30, 2022, to April 30, 2023. Analysis of the data revealed that 40% of the sentiments expressed were positive, 51% were neutral, and the remaining 9% were negative. The findings indicate the overall sentiment analysis, as illustrated in Fig. 2.

Fig. 2
figure 2

Sentiment Analysis: Distribution of Natural, Positive, and Negative Sentiments

These results are similar to previous research regarding the distribution of ChatGPT sentiments. Haque et al. (Fear

The sentiment of fear revealed five codes that can be found in Fig. 7.

Fig. 7
figure 7

Factors Connected with Fear

Job security and plagiarism also were used to describe fear. This may be expected in that the anger regarding ChatGPT possibly causing these issues can then lead to fear, or vice versa. Nonetheless, there were a few nuances of these issues connecting to fear from class concerns to the wider university, such as “Universities fear a cheating epidemic after more than one million visits to ChatGPT during the winter exam season.” Comments such as these also show faculty fears of the reliability of tests in higher education and described fear of assessment threats. The research indicates that writing papers are a common form of assessment in determining what knowledge a student has gained from a class (Rawlusyk, 2018). The ability of ChatGPT to take faculty assignment descriptions and produce papers is a valid fear. ChatGPT can even write in a certain style if asked to do so and can mimic an undergraduate/graduate expected output. Faculty can have great difficulty deciphering what text is created by humans or by AI. Faculty described having to consider new forms of assessment where students are less likely to use ChatGPT to cheat on tests.

As the Internet first emerged, there was considerable concern about students copying and pasting information from the Internet (Garg & Goel, 2023). This appears more of a concern with ChatGPT as it is harder to trace where the program has gathered the information from as it synthesizes text and makes new collections of text. Faculty fears of disruption were often connected to students using ChatGPT to cheat on assessments, by often handing in the work by AI claimed as one’s own work. Removal of these common traditional practices appears to be causing fear with tweets such as “Some educators see ChatGPT as a sign that AI will soon lead to the demise of the academic essay, a crucial tool used in schools at every level.”

Disgust

Disinformation and bias brought a tone of disgust to faculty tweets. Of the negative feelings that faculty have towards ChatGPT, bias only appeared in the disgust sentiment. Disgust comments include, “Maybe the scariest thing about the artificial intelligence writing program, #ChatGPT, is that it has a leftist bias built in”. Bias has been a concern across the literature with a variety of empirical findings showing political bias favoring liberal politicians to conservatives (McGee, 2023). Chen et al. (2023) examined the different bias aspects of ChatGPT and found that as it was pre-trained with human data, ChatGPT has many learned behavioral biases, including (1) conjunction bias, (2) probability weighting, (3) overconfidence, (4) framing, (5) anticipated regret, (6) reference dependence, and (7) confirmation bias (Fig. 8).

Fig. 8
figure 8

Factors Connected with Disgust

Plagiarism appeared yet again in the disgust code. However, this was turned to disgust regarding not just the AI program, but on students’ abilities to cheat. A faculty member wrote, “Students who are otherwise mediocre can show brilliance when it comes to finding ways to cheat, and ChatGPT will undoubtedly be harnessed by students to cheat.” Literature reports a variety of creative and imaginative ways that students are using ChatGPT to cheat (e.g., King & ChatGPT, 2023). The study that King wrote with ChatGPT notes that students can add spelling mistakes to AI text to fool programs into thinking that the text was written by a human. Students could also use the abilities of ChatGPT to write in the tone and format of what an undergraduate or graduate student would with a prompt of “Act as an undergraduate student and write a ….”. This would tell ChatGPT to write in the style of a student.

Threats to writing were similar to issues discussed in the prior negative sentiments. Feelings and beliefs of bias, disinformation, and threats to curriculum content brought comments of disgust, such as “ChatGPT is not yet fit to do the things people are asking it to do; that’s why we’re having an adverse reaction to it.” This sentiment is interesting as ChatGPT is recognized by the majority of tweets as having extensive capabilities that are used for both positive and negative impacts.

Sadness

The grounded coding revealed four factors connected with sadness, shown in Fig. 9. They suggest a sense of concern about ChatGPT, the potential for disinformation, tools associated with AI, job security, and the future.

Fig. 9
figure 9

Factors Connected with Sadness

Similar to anger and disgust, threats to writing and plagiarism were emerging themes for the sadness sentiment. Faculty expressed sadness about plagiarism and the power of ChatGPT to mimic human writing, predicting that AI will lead to the demise of writing. These findings support other studies where faculty suggest that the use of AI signifies the decline of critical writing skills and critical thinking (Amani et al., 2023).

Surprise

The grounded coding revealed three factors connected with surprise, shown in Fig. 11.

Fig. 11
figure 11

Factors Connected with Surprise

The surprise sentiment seems to fit in positive, negative, and neutral categories. A positive surprise sentiment can be found in tweets about ChatGPT’s capabilities and faculty desire to learn more about them. Many of the tweets seemed encouraged about the potential of ChatGPT to provide extraordinary opportunities for academia, including new tools for teaching and learning, or predicted it as a catalyst for transforming the way we assess learning. At times tweets expressed an intense surprise, “Omg I think #chatgpt just blew my mind”. These reactions resonate with other sentiment studies (Taecharungroj, 2023) and reaffirm affordances suggested by advocates of ChatGPT (Bozkurt et al., 2023; Dai et al., 2023).

Other tweets in the surprise sentiment fall in the neutral or negative categories. One tweet expressed amazement at the capabilities and tools provided to students, but with a tone of fear: “It’s scary how good it is.” Another tweet recognized the need to embrace extraordinary opportunities but suggested that it will likely take place “amid the panic” of AI revolutionizing higher education. These tweets expressed positive surprise with the addition of concerns, a pattern also found by Leiter et al. (2022; Rae, 2022). Joy sentiments in this study appear to respond to these appeals, suggesting that faculty are joyful about the potential for AI to, at least partially, meet these challenges. One tweet suggested that the introduction of ChatGPT may be a catalyst to eliminate conventional grading and “implement more authentic and humane ways to document learning”. Other tweets concurred that ChatGPT could invigorate our current teaching strategies by “breaking away from the traditional essay” and introducing creative approaches to evaluation and student engagement, including incorporating effective use of AI in future career contexts.

Similar to the trust sentiment, some tweets in the joy sentiment proposed embracing technology as a vehicle for promoting discussions on academic integrity. Other tweets expressed the value of learning processes by highlighting ChatGPT’s potential to enhance learning for students with special needs, a finding that supports other studies in which AI affordances connected to accessibility and inclusion are suggested (Bozkurt et al., 2023; Dai et al., 2023).

While faculty have many valid concerns about the adoption of ChatGPT in higher education, the tweets found in the joy sentiment suggest that a large number of stakeholders in higher education recognize the promised affordances of ChatGPT and view it in a positive light.

Implications

This study’s findings hold relevance for both higher education stakeholders and policymakers. The research offers faculty a nuanced understanding of their attitudes toward ChatGPT and AI technologies in educational settings. Recognizing the predominantly positive sentiment toward ChatGPT can encourage educators to explore its potential for fostering creativity, enhancing student engagement, and improving accessibility in their teaching practices. This fosters a more informed and responsible integration of AI technologies into their teaching methodologies. Additionally, identifying specific areas where faculty perceive potential benefits of ChatGPT informs future research and development endeavors in educational technology. Furthermore, understanding the spectrum of emotions, including concerns about plagiarism, ethical implications, and job security, empowers faculty to address these issues proactively.

Policymakers can leverage this information to formulate policies and guidelines that support the ethical and effective integration of AI tools like ChatGPT into educational institutions. By addressing plagiarism and job security, policymakers can create an environment that fosters innovation and technology adoption while safeguarding academic integrity and faculty well-being. This insight guides efforts to optimize AI technologies to better align with the needs and objectives of faculty and students, ultimately enhancing teaching and learning outcomes. By aligning the development of AI technologies with faculty needs and preferences, educational institutions can foster an environment conducive to innovation and pedagogical excellence.

Limitations and Future Research

As this study only shows a snapshot of information from X, the findings from this study cannot be generalized to all higher education faculty. Furthermore, only tweets written in English were included which bias the findings towards those countries and contexts that use English. To achieve broader generalizability, future research should explore other platforms such as LinkedIn, Facebook, Instagram, YouTube, and TikTok and incorporate surveys and interviews in diverse languages. It is important to note that this study was conducted within a 5-month timeframe of ChatGPT’s release. This provides an important emerging reaction to ChatGPT during this initial period, but this also limits the understanding of faculty perceptions over a longer duration. Finally, while the VADER sentiment analysis was utilized, based on lexicon-based and rule-based approaches, it is important to acknowledge that the accuracy of dictionary-based methods can vary depending on the characteristics and quality of the data. Future researchers can explore alternative methods such as AFINN, SenticNet, Bing Liu Lexicon, SentiStrength, Opinion Lexicon, or develop machine learning models to analyze sentiment in textual data.

Conclusion

This study explored higher education faculty perceptions of ChatGPT and investigated the factors causing these perceptions. Although previous studies have acknowledged the significance of ChatGPT in the education sector, they have yet to focus on exploring the perception of ChatGPT among faculty members. X data was gathered and analyzed using lexicon-based sentiment analysis and ground coding techniques. The study found positive sentiments were more prevalent in the tweets than negative sentiments. The study also revealed the diverse range of emotions held by higher education faculty regarding ChatGPT, with trust and joy being the most prevalent positive sentiments and fear and anger being the most prevalent negative sentiments. Faculty members generally perceive ChatGPT as a positive advancement in AI technology, recognizing its potential in areas such as creativity, student engagement, optimal learning, accessibility, and inclusion. However, it is vital to acknowledge that concerns were expressed regarding issues such as plagiarism, ethical considerations, and job security. In conclusion, this study provides insights into faculty perceptions of ChatGPT which can aid in better understanding how best to support faculty in incorporating this technology into their practice.