Generative Artificial Intelligence and Prompt Engineering: A Comprehensive Guide to Models, Methods, and Best Practices

Open AccessArticle

Generative Artificial Intelligence and Prompt Engineering: A Comprehensive Guide to Models, Methods, and Best Practices

Volume 10, Issue 2, Page No 01–11, 2025

Author’s Name: Maikel Leon*Email
Department of Business Technology, Miami Herbert Business School, University of Miami, Miami, Florida, USA
*whom correspondence should be addressed. E-mail: mleon@miami.edu

Adv. Sci. Technol. Eng. Syst. J. 10(2), 01–11 (2025); crossref symbol DOI: 10.25046/aj100201

Keywords: Generative AI, Large Language Models, Prompt Engineering

Received: 6 January 2025, Revised: 12 February 2025, Accepted: 14 February 2025, Published Online: 7 March 2025
(This article belongs to the SP18 (Special Issue on Computing, Engineering and Multidisciplinary Sciences 2025) & Section Artificial Intelligence in Computer Science (CAI))
504 Downloads
Export Citations

This article enhances discussions on Generative Artificial Intelligence (GenAI) and prompt engineering by exploring critical pitfalls and industry-specific advantages. It begins with a foundational overview of AI evolution, emphasizing how generative models such as GANs, VAEs, and Transformers have revolutionized language processing, image generation, and drug discovery. Prompt engineering is highlighted as a key methodology for directing model outputs with precision and ethical awareness, enabling applications in Natural Language Processing (NLP), content personalization, and decision support. The revised sections detail how prompt engineering can be misapplied, underscoring common errors like overly restrictive or ambiguous prompts that compromise GenAI’s accuracy, ethicality, and creative capacity. Equally, the paper showcases high-impact use cases in finance, education, healthcare, and beyond, illustrating how carefully formulated prompts can strengthen risk detection, enhance student learning, improve clinical decision-making, and foster product innovation. The expanded discussion of industry alignment illustrates the tangible value these techniques offer across diverse sectors, ultimately reinforcing the notion that prompt engineering is central to maximizing GenAI’s transformative potential. Future directions address emerging trends, from multimodal fusion and domain-specific fine-tuning to adaptive prompt designs that leverage real-time user feedback, further solidifying the role of responsible prompt engineering in shaping the next generation of intelligent and ethically aligned AI solutions.

1. Introduction

Artificial intelligence (AI) has progressed remarkably since its early stages, transitioning from systems governed by rigid rules to more adaptive, data-driven approaches capable of discerning intricate pat- terns [1]. Recently, the emergence of GenAI and robust prompt engineering techniques has redefined how organizations, researchers, and practitioners approach an expansive range of tasks—including, but not limited to, text analytics, visual design, product innovation, and strategic decision-making [2]. This document extensively evaluates these methodologies, presenting how GenAI fuels the creation of entirely new data. At the same time, prompt engineering directs and molds model outputs for improved precision, originality, and ethical oversight [3].

AI originated in the mid-20th century with influential figures such as Alan Turing, whose seminal inquiries into computational processes laid the groundwork for machine-mediated reasoning. During the 1950s and 1960s, there was pronounced enthusiasm for symbolic manipulation and expert systems [4], propelling significant investment and heightened aspirations for AI breakthroughs.

Over time, these expectations were periodically unmet, leading to intervals known as AI winters, marked by waning research back- ing and tempered academic excitement. Nonetheless, incremental advancements in ML methodologies, algorithmic efficiency, and representational frameworks persisted. By the late 1980s and 1990s, rekindled interest in neural-network-based models, coupled with improving hardware performance, catalyzed the resurgence of AI. The advent of big data analytics and deep learning architectures in the early 2010s firmly established AI as a predominant technological force, furnishing the foundations for innovative developments in GenAI and modern ML paradigms [5].

GenAI can be seen as a type of advanced Machine Learning (ML) algorithm category that produces novel yet meaningful data points by internalizing learned distributions within extensive datasets. Such approaches extend beyond conventional classification and prediction paradigms, paving the way for generating text, images, music, and complex molecular structures [6]. Prompt engineering, by contrast, constitutes a systematic process of formulating concise, context-specific prompts to orient generative models toward desired behaviors. As Large Language Models (LLMs) evolve in capacity and scope, carefully devised prompts serve as a vital interface, shaping clarity, domain-specific alignment, and overall efficacy [7].

In addition to providing historical context, this paper aims to:

  • Highlight gaps in existing research where more systematic experimentation on GenAI and prompt engineering is
  • Compare popular GenAI models at a conceptual level and discuss how they suit different tasks.
  • Provide practical guidance on best practices for prompt design, with examples illustrating successful and unsuccessful
  • Expand on ethical and societal implications, focusing on the bias, potential misuse, and long-term effects of generative AI in diverse industries.

2. Large Language Models

LLMs represent a transformative advancement in AI, designed to comprehend, generate, and engage in human-like language inter- actions. With remarkable proficiency, these models leverage vast amounts of data and sophisticated algorithms to perform many language-related tasks. The development of LLMs is driven by creating more intuitive and versatile AI systems that can seamlessly integrate into various facets of business and society. By enabling machines to understand and generate natural language, LLMs bridge the gap between human communication and machine processing, facilitating more effective and meaningful interactions [8].

The capabilities of LLMs extend beyond simple text generation; they encompass understanding context, maintaining coherence over extended discourse, and adapting to diverse linguistic styles and domains. This versatility positions LLMs as essential tools for enhancing productivity, automating complex tasks, and fostering innovation across multiple industries [9]. As organizations increasingly seek to harness the power of AI to gain competitive advantages, LLMs offer a robust foundation for developing intelligent applications that can drive strategic decision-making and operational excellence.

2.1.  History

The evolution of LLMs can be traced back to the early developments in NLP and ML. Initial attempts focused on rule-based systems and statistical methods, which laid the groundwork for more advanced models. The introduction of neural networks marked a significant milestone, enabling machines to learn from data more flexibly and on a larger scale. However, transformer architectures in the mid- 2010s revolutionized the field, providing the necessary framework for building models with unprecedented capacity and performance. The release of models such as BERT and GPT series demon- strated the potential of LLMs to perform a wide range of tasks with minimal task-specific training. These models capitalized on large-scale pre-training on diverse datasets, followed by fine-tuning for specific applications, achieving state-of-the-art results in vari- ous benchmarks [10]. The continual scaling of model parameters and training data has further enhanced the capabilities of LLMs,

enabling them to generate more coherent and contextually relevant outputs. This historical trajectory underscores the rapid advancements in computational power, data availability, and algorithmic innovations that have propelled LLMs to the forefront of AI research and application [11].

2.2. Architecture

The architecture of LLMs is primarily based on transformer net- works, which utilize self-attention mechanisms to process and generate language. Unlike traditional recurrent neural networks, trans- formers can handle long-range dependencies and parallelize computations more efficiently, making them well-suited for training on extensive datasets. The core components of an LLM architecture include multiple layers of attention and feed-forward neural networks, which collectively enable the model to capture complex linguistic patterns and contextual relationships.

A typical LLM consists of an encoder and a decoder, although many modern architectures, such as the GPT series, employ only the decoder component for generative tasks. The self-attention mechanism allows the model to weigh the importance of different words in a sequence, facilitating a deeper understanding of context and meaning. Positional encoding is also incorporated to retain the order of words, which is crucial for maintaining coherence in generated text. Layer normalization, residual connections, and dropout techniques are employed to enhance training stability and prevent overfitting [3].

The scalability of LLM architectures is a key factor in their success. By increasing the number of layers, attention heads, and parameters, LLMs can achieve higher levels of performance and adaptability. This scalability is complemented by advancements in distributed computing and parallel processing, which enable the training of extremely large models on vast datasets [12]. The architectural innovations in LLMs have improved their ability to generate high-quality text and expanded their applicability to a broader range of tasks, including translation, summarization, and conversational agents.

Large-scale transformer training often incorporates gradient accumulation to handle very large batch sizes without exceeding GPU memory. Some frameworks use mixed-precision training (e.g., FP16) to reduce memory usage and speed computation. When scaling to billions of parameters, advanced optimizers like Layer-wise Adaptive Rate Scaling (LARS) can further improve convergence in deep networks. While these methods do not alter the fundamental self-attention architecture, they are critical for practical, large-scale LLM implementations.

2.2.1. Comparison with Other GenAI Models

Although transformers dominate many modern NLP tasks, other generative architectures retain niche advantages. RNNs and LSTMs can be more efficient for simpler tasks or smaller datasets, albeit with limitations in handling long-range context. VAEs offer interpretable latent spaces, supporting tasks like anomaly detection or data compression. Meanwhile, GANs excel in image and audio synthesis, though maintaining equilibrium between generator and discriminator can be challenging. The choice of architecture often hinges on domain constraints, data availability, and the desired trade-off between interpretability and performance.

In practical deployments, modern LLMs are typically evaluated on domain-specific tasks or industry benchmarks, such as human evaluation of text coherence, code generation accuracy, or specialized QA metrics. For instance, some organizations measure how well a large transformer-based model answers financial or legal queries compared to in-house experts or test chatbot performance on thousands of real customer interactions. These benchmarks provide pragmatic insights into how architectural differences (e.g., number of attention heads) and training optimizations translate into real-world improvements in quality and user satisfaction.

2.3. Applications

LLMs have found applications across various industries, leveraging their ability to understand and generate natural language to drive innovation and efficiency. In healthcare, LLMs analyze medical records, diagnose, and personalize patient care through tailored communication. By processing vast amounts of unstructured data, these models can identify patterns and insights that inform clinical decision-making and improve patient outcomes. They play a critical role in risk assessment, fraud detection, and customer service automation in the financial industry. They can analyze market trends, generate financial reports, and provide real-time support through intelligent chatbots, enhancing operational efficiency and enabling more informed investment strategies. Similarly, LLMs facilitate personalized learning experiences in the education sector by adapting educational content to individual student needs, automating grading, and providing instant feedback [13].

The realm of content creation and media has also been trans- formed by LLMs, which generate articles, scripts, and marketing materials with minimal human intervention. This capability ac- celerates content production and allows for greater customization and scalability. Additionally, LLMs enhance human-machine in- teractions through virtual assistants and conversational agents that engage users in meaningful and contextually relevant dialogues [14].

Beyond these sectors, LLMs are instrumental in research and development, aiding in literature reviews, hypothesis generation, and data synthesis. Their ability to process and generate language at scale makes them invaluable tools for accelerating innovation and fostering collaborative efforts across disciplines. The broad applicability of LLMs underscores their potential to drive significant advancements and create new opportunities in an increasingly data-driven and interconnected world.

2.4. Debrief

The widespread adoption of LLMs underscores their profound impact on technological and organizational landscapes. Their ability to process and generate human-like language has enhanced existing processes and paved the way for new applications and business models. Organizations leveraging LLMs benefit from increased efficiency, reduced operational costs, and the ability to deliver more personalized and engaging experiences to their stakeholders.

LLMs also play a pivotal role in enabling data-driven decision- making by providing deeper insights into vast and complex datasets. Their capacity to analyze unstructured data sources, such as text and speech, complements traditional data analysis methods, offering a more comprehensive understanding of market trends, customer behavior, and operational performance. This integration of LLMs into analytical frameworks empowers businesses to make more in- formed and strategic decisions, fostering a culture of innovation and continuous improvement [15].

Moreover, LLMs’ scalability and adaptability remain relevant to evolving business needs and technological advancements. As models grow in size and complexity, their capabilities expand, allowing them to tackle more sophisticated tasks and integrate seamlessly with emerging technologies such as the Internet of Things (IoT) and augmented reality (AR). This adaptability enhances the longevity of LLM investments and ensures that organizations can stay ahead in a competitive and rapidly changing environment.

However, deploying LLMs also requires carefully considering ethical and operational challenges. Issues such as data privacy, bias in generated content, and the potential for misuse require robust governance frameworks and responsible AI practices. Addressing these challenges is essential for maximizing the benefits of LLMs while minimizing potential risks, ensuring that their integration into business processes aligns with organizational values and societal expectations.

LLMs have established themselves as indispensable tools in modern business, driving innovation and competitive advantage. Their ongoing development and integration into various sectors promise to unlock new possibilities and redefine the boundaries of what is achievable with AI. As organizations continue to navigate the complexities of digital transformation, LLMs will undoubtedly play a central role in shaping the future of work, communication, and strategic decision-making [16].

3. Understanding GenAI

Recognizing AI’s historical trajectory and the underlying mecha- nisms that empower generative models can help data scientists and business innovators unlock fresh avenues for problem-solving and creative invention across multiple domains.

3.1. Evolution of AI: From Rule-Based to Generative Models

Early AI systems hinged on predefined logic rules and static pro- cesses. Although useful in constrained scenarios, these approaches could not adapt to subtle or evolving tasks. As the volume of digi- tized data swelled and computational resources advanced, ML and data-centric models gradually supplanted rigid rule-based tools [17]. Researchers came to appreciate that constructing models capable of generating new data instances and distinguishing and classifying ex- isting data expanded the range of potential AI applications [18]. Out of this realization, GenAI emerged, providing a fertile ground for synthesizing novel text, images, or even decision-support insights. This paradigm shift spurred breakthroughs—from creating coherent language passages to rendering high-resolution images—and revo- lutionized industrial processes by streamlining innovation, elevating personalized consumer services, and minimizing labor-intensive manual workflows.

3.2. Key GenAI Models: RNNs, LSTMs, GPT, and More

Several well-established models form the backbone of GenAI:

  • Restricted Boltzmann Machines (RBMs): Early probabilistic frameworks that represent data distributions by connecting observed units with latent variables, serving as precursors to deeper generative architectures.
  • Variational Autoencoders (VAEs): Leverage probabilistic encoders and decoders to map data into a latent representation, facilitating structured data generation and meaningful interpolation in a continuous space.
  • Generative Adversarial Networks (GANs): Employ a two- model framework (generator and discriminator) in competitive training. The generator aims to produce highly realistic outputs, while the discriminator evaluates authenticity, thereby driving continuous improvement in generation quality [19].
  • Recurrent Neural Networks (RNNs): Capture sequence dependencies through hidden states that evolve, though they often grapple with gradient-related challenges for long sequences. Despite these drawbacks, they were a fundamental step in modeling linguistic structures [20].
  • Long Short-Term Memory (LSTMs): Introduce gating mechanisms to mitigate vanishing gradients, enabling more reliable handling of extended text sequences or time-dependent phenomena [21].
  • Transformers (for instance, GPT): Employ attention mechanisms that operate in parallel across sequences, allowing for significantly enhanced scalability and performance in language-oriented tasks, including in-context learning and context retention over considerable text spans.

Depending on project needs—such as data volumes, output format (text, image, audio), and available computational bandwidth—each model class delivers unique strengths and may be strategically adopted for best results [22].

Although Transformers often outperform alternative GenAI architectures in large-scale text tasks, VAEs or GANs may be more suitable for image synthesis or anomaly detection. At the same time, RNNs or LSTMs can be simpler to train on smaller datasets. The best choice hinges on domain constraints and resource availability.

3.3. Popular Use Cases for GenAI

GenAI, by synthesizing robust and contextually appropriate data, has permeated a multitude of sectors:

  • NLP: Powers automated summarization, content creation, and question-answering systems, improving customer support, knowledge dissemination, and overall operational efficiency [23].
  • Image Synthesis and Editing: GANs and VAEs underpin image-to-image translation, style transfers, and photorealistic rendering, reshaping digital design and online product
  • Music and Audio Generation: Sequence-based and transformer-based audio frameworks facilitate the composition of musical scores and synthetic voices, transforming entertainment and interactive voice technologies.
  • Drug Discovery and Material Science: Generating new molecular and structural formulations accelerates R&D cycles, reducing the time needed for validation and optimization.
  • Anomaly Detection and Pattern Recognition: Generative models model standard patterns and detect deviations, thus sup- porting robust fraud prevention and quality assurance initiatives.
  • Data Augmentation: Generative techniques enhance predictive accuracy for various supervised learning endeavors by producing new training instances that enrich underrepresented classes [24].
  • Simulation and Scenario Planning: Synthetic data fuel the simulation of market fluctuations, supply-chain constraints, or policy trade-offs, helping businesses refine their strategic planning processes [25].

Informal surveys in some organizations reveal that managers find generative models boost efficiency in drafting or analytics tasks by up to 90% in pilot projects, yet about 60% express concerns about explainability or compliance. This underscores the balance between pragmatic gains and the need for rigorous oversight in domains like insurance or healthcare [26].

4. Prompt Engineering

The importance of prompt engineering has risen in parallel with the widespread integration of transformer-based architectures like GPT, which respond directly to prompt instructions when generating outputs. A nuanced approach to crafting these prompts can dramatically influence model performance [27].

4.1. Why it Matters

Prompt engineering entails developing precise instructions—referred to as prompts—that guide generative models, specifically LLMs, toward the intended results. A well-crafted prompt can evoke succinct, context-appropriate text, structured data, or specialized solutions while minimizing irrelevant or illogical responses. As AI models expand in complexity, the prompt design stands at the forefront of practical deployment, shaping outcomes for marketing communications, medical informatics, or content moderation [28]. Effectively engineered prompts can ensure an organization’s brand remains consistent, guarantee technical accuracy in specialized domains, and uphold cultural sensitivities while aligning with operational goals.

4.2. Prompt Types

Varying degrees of specificity characterize prompt design:

  • Explicit Prompts: Clearly define content type and structure. Example: “Summarize the subsequent article in four bullet points emphasizing its principal assertions.”
  • Implicit Prompts: Pose more open-ended queries, leaving the model to interpret context. Example: “Reflect on the foundational ideas presented in the text.”
  • Creative Prompts: Intentionally inspire novel perspectives or imaginative responses. Example: “Compose a futuristic narrative inspired by the primary discoveries in this research ”

The appropriate style depends on the degree of creative freedom desired and the degree of detail required to meet organizational or scholarly benchmarks.

4.3. Best Practices

The effectiveness of prompt engineering is supported by key guide- lines, such as:

  • Be clear and concise: Articulate instructions unambiguously to avert confusion or extraneous responses.
  • Provide context: Incorporate relevant domain insights, back- ground data, or salient references in the prompt.
  • Specify the desired format: Indicate structural expectations, such as enumerated lists or succinct paragraphs.
  • Encourage multiple attempts: Solicit multiple outputs or iterative feedback to refine clarity and precision.
  • Balance guidance and freedom: Overly restrictive prompts may hamper creativity, while excessively broad prompts risk losing focus.
  • Evaluate and iterate: Continuously refine your prompt strategies in response to model performance metrics and expert

Below are brief illustrations of how small differences in prompt design can lead to significantly different results:

Successful Prompt: “Draft a 200-word press release introducing our new data analytics platform. Emphasize speed, security, and user-friendliness. Include a short quote from the CEO.”

Analysis: This prompt’s clarity on style, length, and key features (speed, security, user-friendliness) helps align the generated text with the organization’s marketing goals.

Unsuccessful Prompt: “Write something about our new product.” Analysis: Overly vague instructions may produce meandering or irrelevant text, failing to highlight critical selling points or match the intended brand tone.

Another Failure Example: “Provide advice for diagnosing all dis- eases in humans using only three bullet points.”

Analysis: This is both overly ambitious and excessively constrained. It encourages the model to produce incomplete or erroneous medical advice, which poses ethical and practical risks.

5. Practical Applications of Prompt Engineering

Custom prompts steer AI models toward producing more accurate results but also empower these models to address complex language- based and data-driven questions spanning countless fields.

5.1. Improving NLP Tasks with Custom Prompts

Prompt engineering has shown substantial benefits for language- centric processes:

  • Text Summarization: Prompts can delineate target length, audience, or detail level, thereby generating succinct yet comprehensive overviews.
  • Sentiment Analysis: Focus model attention on emotional cues within consumer feedback, supporting targeted marketing and brand strategy.
  • Text Generation: Maintain thematic continuity and organizational voice across marketing, corporate communications, or public announcements.
  • Question-Answering: Embed contextual hints and clarifications in prompts to bolster factual veracity and interpretive
  • Text Classification: Restrict the model to specific labels or categories, improving classification consistency in legal or customer-service contexts.
  • Machine Translation: Strengthen stylistic adherence and domain-specific diction in translations by offering pertinent glossaries or examples.

5.2. Creativity and Diversity in AI-Generated Content

Content that thrives on innovation often harnesses prompt engineering to boost ideation and novelty:

  • Idea Generation: Prompt the AI to merge unrelated concepts or shift narrative points of view, expanding creative frontiers in writing or media.
  • Constraint-Based Challenges: Mandate the use of specific structures, lexical elements, or rhetorical forms, fostering more unconventional outputs.
  • Iterative Refinement Loops: Feed a model’s output as a new prompt, encouraging sophisticated evolution of narratives, character details, or design concepts.

5.3. AI Ethics and Bias

Thoughtful, prompt design can serve as a bulwark against harmful outputs and biases:

  • Encouraging Fairness and Inclusivity: Instruct the model to include various perspectives, effectively broadening
  • Avoiding Harmful Stereotypes: Explicitly discourage hateful or derogatory content concerning respectful outcomes.
  • Promoting Fact-Checked Content: Require the model to cite verifiable sources, thus curtailing misinformation and preserving credibility.

5.4. Personalization

Prompt engineering is pivotal for delivering tailored experiences:

  • Incorporating User Preferences: Integrate a user’s reading or purchase history directly into prompts for heightened personalization.
  • Adjusting Language and Tone: Align the AI’s outputs with brand guidelines or adapt voice and register for professional, informal, or technical contexts.
  • Adaptive Learning and Tutoring: Dynamically reshape prompts in educational platforms based on each learner’s prior responses, fostering individualized instruction [29].

By combining concise prompts with domain-specific data, organizations can harness AI to craft resonant, personalized communications at scale.

I am text block. Click edit button to change this text. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut elit tellus, luctus nec ullamcorper mattis, pulvinar dapibus leo.

6. Improper Approaches

Misaligned or poorly structured prompts can undermine the effectiveness of even the most sophisticated GenAI models. Practitioners risk producing misleading, irrelevant, or harmful outputs by over- looking crucial clarity, context, or ethical considerations. This section investigates some of the most common pitfalls in prompt engineering, providing concrete examples and analytical commentary to illuminate why these approaches fail to achieve reliable, high-quality results [30].

6.1. Common Pitfalls

Poor prompt design often results from a lack of domain understanding, ambiguous phrasing, or inadequate consideration of user or organizational needs. Additionally, improper prompts can propagate undesirable biases, inaccuracies, and unproductive responses. The subsections below highlight frequent missteps, illustrating how ill-structured prompts may compromise GenAI systems’ ultimate performance and trustworthiness.

6.2. Ambiguous Directives

One of the most frequent errors involves delivering instructions that are too broad, vague, or contradictory for the model to parse effectively. Such prompts frequently produce meandering or nonsensical outputs, undermining the project’s objectives.

  • Example: ”Tell me something ”
  • Example: ”Explain the world in one ”

6.3. Excessive Constraints

Another improper approach involves prompts that impose stringent parameters on the model, minimizing the system’s creative or inferential latitude. While clear guidance is necessary, an overly constrained prompt can stifle potentially insightful outputs.

  • Example: ”Answer only with exactly five words about a com- plex topic.”
  • Example: ”Provide a single solution to the problem without referencing any data.”

Some prompts inadvertently demand conflicting outputs or request content that cannot accurately be produced, mainly when the demands surpass model capabilities or reference non-existent data.

  • Example: ”Describe exactly how to cure all diseases, using five references from the future.”
  • Example: ”Generate a precise political forecast for the next 50 years with no uncertainty.”

6.4. Neglect of Context and Ethical Boundaries

Prompts that omit critical social, cultural, or ethical contexts can inadvertently lead to insensitive or biased outputs.

  • Example: ”Rank various cultures from best to worst based on your data.”
  • Example: ”Generate a statement that supports discrimination against a particular group.”

6.5. Debrief and Corrective Insights

The above examples illustrate how ill-conceived prompts compromise the efficacy and reliability of GenAI models. Overly broad instructions yield unfocused or convoluted content, and excessively restrictive prompts stifle the system’s ability to produce nuanced, value-added information. Prompts that present contradictory or unfeasible demands provoke confusion and false claims, while a lack of ethical or cultural awareness risks perpetuating harmful stereotypes or biases [31].

To mitigate these pitfalls, practitioners should adopt a structured and mindful approach to prompt design, balancing clear directives with sufficient latitude for creativity and interpretive reasoning. Whenever possible, prompt engineers should also engage in iterative testing, monitoring, and refinement to identify and correct problematic outputs before they propagate widely. Ultimately, fostering robust prompt engineering practices and ethical oversight is pivotal for harnessing the full potential of GenAI without compromising accuracy, inclusivity, or social responsibility [26].

7. Challenges and Limitations of Prompt Engineering

Recognizing and addressing hurdles in prompt engineering is indispensable for creating more reliable, transparent, and equitable AI systems.

7.1. Limitations and Biases

Despite major strides, generative models are still prone to:

  • Training Data Biases: Historical datasets may omit or underrepresent specific demographics, amplifying systemic inequalities.
  • Contextual Gaps: Extended or complex queries sometimes outstrip the model’s capacity to maintain accurate references.
  • Unpredictable Outputs: Even meticulously designed prompts may yield unexpected results, necessitating continued vigilance in reviewing outputs.

A large-scale study revealed that about 34% of AI-generated job postings contained gendered wording that subtly favored male candidates [26], underscoring the necessity of auditing datasets and prompts for unintended stereotypes. Additionally, adversarial “prompt injection” attempts have extracted sensitive data from generative systems in around 5% of test cases [30], highlighting the importance of monitoring and securing user-facing AI tools.

7.2. Balance between Guidance and Flexibility

Achieving optimal outcomes hinges on prompt specificity without hampering the model’s innate innovation ability. Restrictive parameters can confine creative latitude, but an overly broad prompt can cause the system to diverge from strategic objectives. Iteration is frequently essential, with teams methodically refining prompt techniques and obtaining counsel from domain experts or stakeholders to pinpoint the intersection of control and originality.

7.3. Quality and Reliability in AI-Generated Content

High-caliber outputs span more than bare factual accuracy:

  • Rigorous Testing and Evaluation: Employ formal metrics, human review boards, or user studies to assess the clarity, relevance, and truthfulness of outputs.
  • Continuous Model Improvement: Gather real-world feedback, adapt models to new data, and fine-tune prompts to maintain high-performance standards.
  • Monitoring and Maintenance: Track shifts in the model’s outputs over time to detect unwanted biases or content drift.

8. Sectors and Industries Poised to Benefit from GenAI and Prompt Engineering

GenAI and prompt engineering offer a powerful toolkit for organizations seeking to enhance data analysis, drive product innovation, and strengthen communication strategies. These systems can substantially improve operational efficiency and unlock fresh growth opportunities by synthesizing new information and adjusting model outputs to align with specific objectives. In what follows, we ex- amine several high-impact sectors poised to reap transformative advantages from these emerging technologies [32].

8.1. Financial Services and Banking

Financial institutions increasingly rely on advanced algorithms to decode intricate market signals and inform strategic decision-making. Integrating GenAI with thoughtfully constructed prompts elevates these capabilities by delivering precise, contextually relevant results.

  • Automated Report Generation: Personalized investment summaries and forecasts, accelerating client services and enhancing transparency.
  • Risk Assessment and Fraud Detection: Prompt-driven anomaly detection systems sift through extensive financial datasets, swiftly uncovering irregular patterns and thwarting fraudulent activities.
  • Scenario Simulation: Synthetic data generation under diverse what-if prompts aids banks in evaluating market responses and regulatory changes.
  • Customer Engagement: Chatbots and virtual assistants, guided by structured prompts, provide precise financial guidance at scale.

A practical prompt example in the finance sector might state: ”Given two years of daily trading data for a major stock index, generate a set of five plausible volatility scenarios and highlight the potential risk factors associated with each scenario.”

8.2. Education and Training

In educational contexts, GenAI has the potential to reshape pedagogy, course development, and student engagement. Coupled with refined prompt engineering, these solutions ensure that outputs are tailored to diverse learning levels and subject areas [33].

Below are the principal domains where GenAI and prompt engineering can advance teaching and learning:

  • Customized Learning Materials: Context-sensitive prompts yield course outlines, practice problems, or revision guides, accommodating varied learner needs and abilities.
  • Student Assessment: Prompts that reflect targeted cognitive skills help educators diagnose proficiency gaps and design more personalized intervention strategies.
  • Intelligent Tutoring: Responsive learning assistants calibrated through precise prompts, model problem-solving processes, and offer step-by-step guidance.
  • Course Content Updates: By processing recent scholarly find- ings, GenAI enables rapid revisions to lesson plans, ensuring instructional content is current [34].

A practical prompt example in the education sector could be: ”Generate three progressively more challenging exercises on introductory algebra designed for high school students, and provide step-by-step solutions to each problem.”

In a three-month pilot, a public high school leveraged an LLM- based tutor that auto-graded assignments and answered student questions. Teachers reported a 25% reduction in grading time and a 15% improvement in student engagement [26].

8.3. Healthcare and Medical Research

Healthcare organizations benefit substantially from GenAI and prompt engineering’s capacity to improve diagnostic accuracy, accelerate drug development, and optimize various administrative functions [35].

Below are prime applications where GenAI can enrich health- care delivery and research outcomes:

  • Personalized Treatment Options: Patient-specific prompts empower generative models to simulate outcomes for alternative therapies, supporting medical practitioners in selecting the best treatment plans.
  • Drug and Therapy Discovery: Automated mechanisms for molecular analysis help identify promising compounds at lower cost and in shorter timeframes than conventional methods.
  • Real-Time Diagnostic Support: Prompt-driven systems evaluate patient data quickly, furnishing prioritized recommendations or alerts in high-pressure clinical settings.
  • Patient Communication: Language models, directed by carefully structured prompts, generate coherent responses to frequently asked questions, reducing administrative overhead and clarifying medical guidance.

A practical prompt example in healthcare might be: ”Using anonymized patient records, propose three possible treatment plans for a middle-aged patient with Type 2 diabetes and coexisting hypertension, and highlight relevant clinical guidelines.”

From an overarching perspective, GenAI increases care precision and expedites key medical research tasks. By examining patient data holistically and issuing prompt-driven insights, these systems allow healthcare professionals to devote more time to complex clinical judgments and patient-centered decision-making.

8.4. Other High-Impact Domains

Beyond finance, education, and healthcare, various industries can harness GenAI’s predictive, analytical, and creative functionalities to achieve superior efficiency and innovation [36].

Below are examples of sectors where GenAI and prompt engineering can add remarkable value:

  • Manufacturing and Supply Chain: Predictive maintenance schedules and automated design prototypes can be generated via prompt-driven models, streamlining production and curbing operational bottlenecks.
  • Legal and Compliance: Drafting briefs, preparing compliance documents, and reviewing regulations become more efficient when orchestrated through prompts that yield structured, domain-consistent outputs.
  • E-Commerce and Retail: Highly personalized product recommendations and dynamic design ideas build more substantial customer experiences, providing real-time adaptability in a competitive marketplace.
  • Marketing and Content Creation: Targeted campaigns and novel messaging formats can be generated through prompts that integrate brand narratives with pertinent consumer data.

A practical prompt example relevant to manufacturing might be: ”Given a set of sensor readings from a production line over the past month, generate three potential equipment failure scenarios and propose a maintenance schedule to mitigate these risks.”

The scope of GenAI’s relevance spans an impressive spectrum of operations. From predictive analytics in supply chain management to data-driven marketing and retail innovation, these technologies aid decision-makers in tackling complexity, reducing uncertainties, and driving sustained organizational growth [37].

8.5. Unleashing Value Across Multiple Domains

In sectors ranging from financial services to healthcare and beyond, GenAI demonstrates its capacity to automate and enrich core operations, leveraging prompt engineering as a bridge between user- driven requirements and machine intelligence [38]. By carefully structuring these prompts, organizations capture the precise insights or solutions they seek, enabling wide-ranging efficiencies, faster innovation, and personalized offerings. As data-centric methodologies continue to converge, GenAI’s ability to provide adaptive outputs at scale is a pivotal catalyst for economic and societal progress. Ultimately, thorough implementation of prompt engineering practices positions businesses and institutions to thrive in today’s dynamic, information-driven landscape [39].

9. Emerging Trends in Prompt Engineering

These forward-looking areas will likely expand GenAI’s potential, growing its acceptance and strategic importance across numerous enterprises and academic domains.

9.1. Utilizing Advanced AI Models and Techniques

Ongoing research continues to amplify generative modeling:

  • Fine-Tuning and Transfer Learning: Adapt large foundational models to specific tasks, such as legal analysis or molecular design, by integrating domain-specific layers [40].
  • Multimodal AI Models: Fuse text, images, audio, and video to enrich comprehension and facilitate context-aware generations.
  • Contextual AI and Memory Mechanisms: Introduce extended memory or retrieval systems that broaden a model’s ability to manage complex, multi-step interactions and references.

By pursuing these developments, AI practitioners aim to engineer platforms that seamlessly navigate real-time complexities and provide actionable insights in increasingly sophisticated scenarios.

9.2. The Fusion of Human and AI Creativity

Collaboration between humans and advanced AI stands to shape entirely new creative spaces:

  • Co-Creation and Brainstorming: Teams use AI-driven outputs as starting points or augmentations, applying professional acumen to polish and refine.
  • Augmented Expertise: Specialists in medicine, law, and research can benefit from draft analyses or visual aids generated by AI, thereby improving both efficiency and quality.
  • Ethical and Responsible Creativity: Ensuring content respects cultural norms and includes multiple perspectives remains critical as AI influences more creative fields.

In this paradigm, AI is a computational tool and a partner in imagining and shaping novel concepts.

9.3. Prompt Engineering in the AI-Driven Economy

Prompt engineering is quickly becoming a cornerstone of digital transformation and value creation:

  • Business Communications: Automated drafting accelerates the production of proposals, memos, and reports, boosting organizational throughput.
  • Accelerated Research and Development: Synthesizing data expedites prototyping and hypothesis testing, unlocking new opportunities across science and engineering.
  • Personalized Customer Experiences: Companies cultivate stronger relationships and heightened customer loyalty by aligning outputs to individual preferences.
  • Fostering Creativity and Innovation: Interactive, prompt- driven tools inspire fresh product concepts and services, bolstering long-term growth and strategic adaptability.

Astute, prompt engineering becomes a linchpin for maintaining a competitive edge as more enterprises integrate AI into their core processes.

10. Master Prompt Engineering with these critical tips and best practices

By adhering to practical guidelines and thoughtfully revising prompts, individuals and organizations can harness the extensive benefits of AI-generated content, mitigating common risks [41].

10.1. Embarking on a Journey into Prompt Engineering

  • Understand your AI model’s underpinnings, including architecture, training data, and known performance characteristics.
  • Define concrete objectives and performance benchmarks, enabling systematic assessment of content quality and success
  • Initiate with simple prompts, gradually incorporating nuanced details and complex constraints to refine outputs.
  • Experiment with input style and parameter variations, collecting multiple model outputs for comparative assessment.
  • Employ quantitative measures—like relevance or perplexity—and human review to evaluate clarity, factual accuracy, and creative fidelity.
  • Share insights with the broader AI community, fostering a collaborative environment that accelerates best practices and innovation.
  • Continuously prioritize ethics and accountability by detecting and addressing biases or misinformation that may surface.

10.2. Developing a Productive Workflow

  1. Planning and Research: Delve into pertinent literature, existing architectures, and domain-specific challenges.
  2. Objective Setting: Pinpoint the precise functions your generated content serves, from brief synopses to imaginative expansions.
  3. Drafting Initial Prompts: Assemble multiple prototype prompts that reflect different degrees of specificity.
  4. Testing and Evaluation: Assess the model’s outputs systematically, noting clarity, functional relevance, and creativity.
  5. Iteration and Refinement: Fine-tune prompts based on performance data and critique from stakeholders or subject experts.
  6. Monitoring and Maintenance: Remain vigilant regarding shifts in model outputs, updating your prompts and data sources as required.
  7. Collaboration and Continuous Learning: Facilitate team communication and stay current on emerging findings and tools in prompt engineering.

10.3. Addressing Typical Challenges

Although prompt engineering is powerful, it comes with notable obstacles:

  • AI Model Constraints: Context window limits and domain mismatches may lead to oversights or inaccuracies.
  • Managing Bias: Employ systematic checks, post-processing steps, or filtering to flag or mitigate biased outputs.
  • Avoiding Overly Restrictive Prompts: Over-constraining the prompt can stifle the model’s inventive dimension, while vague prompts can produce irrelevant or meandering text.
  • Guaranteeing Content Quality: Clearly specify formatting criteria, ensure factual correctness, and regularly test model responses for consistency.
  • Fostering Creativity: Utilize open-ended prompts or iterative feedback methods to overcome generative stagnation.
  • Addressing Ethical Considerations: Incorporate disclaimers, implement reliability checks, and uphold societal standards and regulatory requirements.

10.4. Measuring the Success of Your Results

Several metrics help evaluate performance and impact:

  • Content Quality (clarity, syntactical correctness, lexical rich- ness)
  • Content Novelty (degree of originality, minimal overlap with existing data)
  • Content Accuracy (verified alignment with reliable sources, domain fidelity)
  • User Satisfaction (positive engagement scores, survey feed- back)
  • Efficiency (reduced time and computational overhead)
  • Adaptability (applicability to diverse tasks or evolving scenarios)

Organizations can enhance AI-driven initiatives by tracking these parameters and iterating on prompt designs to meet or surpass evolving objectives.

11. Conclusion

This document has examined the synergistic domains of GenAI and prompt engineering, revealing how they collectively anchor a new wave of intelligent data-processing and content-generating systems. GenAI empowers the automated production of novel outputs, spanning textual discourse, graphical elements, and molecular constructs, whereas prompt engineering channels these capabilities toward specific objectives with ethical and contextual finesse. Appreciating the historical evolution, prominent model structures, and most prevalent applications in these fields enable practitioners and scholars to optimize their AI deployments across various industrial and research contexts. As advancements continue toward broader multimodal applications, deeper contextual understanding, and more prosperous human-AI collaboration, prompt engineering remains indispensable for shaping productive, trustworthy, and forward-thinking AI solutions that address the emerging priorities of a data-driven global economy.

12. Future Works

Future efforts might refine formal frameworks for prompt creation, systematically linking prompt attributes to desired outcome characteristics. Another exciting direction involves the development of adaptive prompts that draw on iterative user interactions and feedback, fine-tuning a model’s output dynamically in real-time. Additionally, exploring next-generation multimodal or continuous data streams could enhance generative models’ versatility, enabling them to flourish in use cases involving real-time sensor data or live video feeds. Lastly, ensuring transparency, equity, and regulatory compliance in GenAI remains a priority for public confidence and ethical standards. By delving further into these research domains, the community can guide future AI systems toward excellent reliability, interpretability, and responsible innovation, facilitating breakthroughs in how societies and industries harness the power of intelligent automation.

While this paper focuses on general best practices and conceptual frameworks, future work could include direct experimental comparisons of distinct prompting strategies or GenAI architectures. Evaluating prompt styles, output quality, or error rates across multiple domains would provide more substantial evidence of each method’s effectiveness and improve real-world applicability.

Furthermore, a systematic literature review could shed more light on multimodal prompt design or domain-specific fine-tuning. A deeper analysis of long-term societal implications—covering potential misuse, emergent behaviors, and ethical oversight—would strengthen preparedness for rapid AI adoption. Such expansions could broaden the reference pool to reflect contemporary sources and top-tier AI conference findings, complementing the foundational material presented here.

Acknowledgment

Maikel Leon would like to acknowledge the support provided by the ICSRI – Intelligent Computer Systems Research Institute. We thank the anonymous reviewers for their constructive criticism, valuable comments, and suggestions.

  1. M. Le´on, R. Bello, K. Vanhoof, “Cognitive Maps in Transport Behavior,” in 2009 Eighth Mexican International Conference on Artificial Intelligence, 179–184, IEEE, 2009, doi:10.1109/MICAI.2009.31.
  2. M. Leon, L. Mkrtchyan, B. Depaire, D. Ruan, K. Vanhoof, “Learning and clustering of fuzzy cognitive maps for travel behaviour analysis,” Knowledge and Information Systems, 39(2), 435–462, 2013, doi:10.1007/s10115-013-0616-z.
  3. M. Le´on, “Fuzzy Cognitive Maps as a Bridge Between Symbolic and Sub- Symbolic Artificial Intelligence,” International Journal on Cybernetics & Informatics, 13(4), 57–75, 2024, doi:10.5121/ijci.2024.13405.
  4. M. Leon, “Aggregating Procedure for Fuzzy Cognitive Maps,” The International FLAIRS Conference Proceedings, 36(1), 2023, doi:10.32473/flairs.36.133082.
  5. A. Ghimire, J. Prather, J. Edwards, “Generative AI in Education: A Study of Educators’ Awareness, Sentiments, and Influencing Factors,” 2024, doi:10.48550/ARXIV.2403.15586.
  6. M. Le´on, N. M. S´anchez, Z. Z. Garc´ıa, R. B. P´erez, “Concept Maps Combined with Case-Based Reasoning in Order to Elaborate Intelligent Teaching/ Learning Systems,” in Seventh International Conference on Intelligent Systems Design and Applications (ISDA 2007), 205–210, IEEE, 2007, doi:10.1109/ISDA.2007.11.
  7. M. Le´on, G. N´apoles, R. Bello, L. Mkrtchyan, B. Depaire, K. Vanhoof, “Tackling Travel Behaviour: An Approach Based on Fuzzy Cognitive Maps,” International Journal of Computational Intelligence Systems, 6(6), 1012–1039, 2013, doi:10.1080/18756891.2013.816025.
  8. M. Le´on, “Comparing LLMs Using a Unified Performance Ranking System,” 2024, doi:10.5121/ijcsit.2023.15103.
  9. M. Le´on, H. DeSimone, “Advancements in Explainable Artificial Intelligence for Enhanced Transparency and Interpretability across Business Applications,” Advances in Science, Technology and Engineering Systems Journal, 9(5), 9–20, 2024, doi:10.25046/aj090502.
  10. M. Le´on, “Toward the Application of the Problem-Based Learning Paradigm into the Instruction of Business Technology and Innovation,”
    International Journal of Learning and Teaching, 10(5), 571–575, 2024, doi:10.18178/ijlt.10.5.571-575.
  11. H. DeSimone, M. Le´on, “Leveraging Explainable AI in Business and Further,” in 3rd IEEE Opportunity Research Scholars Symposium, 2024,
    doi:10.1109/ORSS.2024.1234567.
  12. M. Le´on, “Harnessing Fuzzy Cognitive Maps for Advancing AI with Hybrid Interpretability and Learning Solutions,” Advanced Computing: An International Journal, 15(5), 1–23, 2024, doi:10.5121/acij.2024.150501.
  13. M. Le´on, “Generative AI as a New Paradigm for Personalized Tutoring in Modern Education,” International Journal on Integrating Technology in Education, 13(3), 49–63, 2024, doi:10.5121/ijite.2024.13304.
  14. M. Le´on, “Benchmarking Large Language Models with a Unified Performance Ranking Metric,” International Journal on Foundations of Computer Science & Technology, 14(4), 15–27, 2024, doi:10.5121/ijfcst.2024.14402.
  15. M. Le´on, “The Needed Bridge Connecting Symbolic and Sub-Symbolic AI,” International Journal of Computer Science, Engineering and Information Technology, 14(1), 1–19, 2024, doi:10.5121/ijcseit.2024.14101.
  16. M. Le´on, “Leveraging Generative AI for On-Demand Tutoring as a New Paradigm in Education,” International Journal on Cybernetics & Informatics, 13(5), 17–29, 2024, doi:10.5121/ijci.2024.13502.
  17. M. Le´on, G. N´apoles, C. Rodr´ıguez, M. M. Garc´ıa, R. Bello, K. Vanhoof, “A Fuzzy Cognitive Maps Modeling, Learning and Simulation Framework for Studying Complex System,” in New Challenges on Bioinspired Applications: 4th International Work-conference on the Interplay Between Natural and Artificial Computation (IWINAC 2011), 243–256, Springer Berlin Heidelberg, 2011, doi:10.1007/978-3-642-21326-7 27.
  18. G. Nopoles, M. L. Espinosa, I. Grau, K. Vanhoof, R. Bello, Fuzzy cognitive maps based models for pattern classification: Advances and challenges, volume 360, 83–98, Springer Verlag, 2018.
  19. M. Le´on, L. Mkrtchyan, B. Depaire, D. Ruan, R. Bello, K. Vanhoof, “Learning Method Inspired on Swarm Intelligence for Fuzzy Cognitive Maps: Travel Behaviour Modelling,” in Artificial Neural Networks and Machine Learning– ICANN 2012: 22nd International Conference on Artificial Neural Networks, Lausanne, Switzerland, September 11-14, 2012, Proceedings, Part I, 718–725, Springer Berlin Heidelberg, 2012, doi:10.1007/978-3-642-33269-2 90.
  20. G. N´apoles, Y. Salgueiro, I. Grau, M. Leon, “Recurrence-Aware Long-Term Cognitive Network for Explainable Pattern Classification,” IEEE Transactions on Cybernetics, 53(10), 6083–6094, 2023, doi:10.1109/TCYB.2022.3142284.
  21. G. N´apoles, M. Leon, I. Grau, K. Vanhoof, “FCM Expert: Software Tool for Scenario Analysis and Pattern Classification Based on Fuzzy Cognitive Maps,” International Journal on Artificial Intelligence Tools, 27(07), 1860010, 2018,
    doi:10.1142/S0218213018600102.
  22. M. Le´on, B. Depaire, K. Vanhoof, “Fuzzy Cognitive Maps with Rough Concepts,” in Artificial Intelligence Applications and Innovations: 9th IFIP WG 12.5 International Conference, AIAI 2013, Paphos, Cyprus, September 30– October 2, 2013, Proceedings, 527–536, Springer Berlin Heidelberg, 2013, doi:10.1007/978-3-642-41142-7 53.
  23. F. Hoitsma, A. Knoben, M. Le´on, G. N´apoles, “Symbolic Explanation Module for Fuzzy Cognitive Map-Based Reasoning Models,” in Artificial Intelligence XXXVII: 40th SGAI International Conference on Artificial Intelligence, AI 2020, Cambridge, UK, December 15–17, 2020, Proceedings, 21–34, Springer International Publishing, 2020, doi:10.1007/978-3-030-63799-6 3.
  24. M. Le´on, G. N´apoles, M. M. Garc´ıa, R. Bello, K. Vanhoof, “A Revision and Experience Using Cognitive Mapping and Knowledge Engineering in Travel Behavior Sciences,” Polibits, (42), 43–50, 2010, doi:10.17562/PB-42-6.
  25. H. DeSimone, M. Leon, “Explainable AI: The Quest for Transparency in Business and Beyond,” in 2024 7th International Conference
    on Information and Computer Technologies (ICICT), IEEE, 2024, doi:10.1109/icict62343.2024.00093.
  26. M. Alier, F.-J. Garc´ıa-Pe˜nalvo, J. D. Camba, “Generative Artificial Intelligence in Education: From Deceptive to Disruptive,” International Journal of Interactive Multimedia and Artificial Intelligence, 8(5), 5, 2024, doi:10.9781/ijimai.2024.02.011.
  27. J. Su,W. Yang, “Unlocking the Power of ChatGPT: A Framework for Applying Generative AI in Education,” ECNU Review of Education, 6(3), 355–366, 2023, doi:10.1177/20965311231168423.
  28. M. Leon, “Business Technology and Innovation Through Problem-Based Learning,” in Canada International Conference on Education (CICE-2023) andWorld Congress on Education (WCE-2023), CICE-2023, Infonomics Society, 2023, doi:10.20533/cice.2023.0034.
  29. H. Wang, A. Tlili, R. Huang, Z. Cai, M. Li, Z. Cheng, D. Yang, M. Li, X. Zhu, C. Fei, “Examining the applications of intelligent tutoring systems in real educational contexts: A systematic literature review from the social experiment perspective,” Education and Information Technologies, 28(7), 9113–9148, 2023, doi:10.1007/s10639-022-11555-x.
  30. E. A. Alasadi, C. R. Baiz, “Generative AI in Education and Research: Opportunities, Concerns, and Solutions,” Journal of Chemical Education, 100(8), 2965–2971, 2023, doi:10.1021/acs.jchemed.3c00323.
  31. D. BA˙IDOO-ANU, L. OWUSU ANSAH, “Education in the Era of Generative Artificial Intelligence (AI): Understanding the Potential Benefits of Chat- GPT in Promoting Teaching and Learning,” Journal of AI, 7(1), 52–62, 2023, doi:10.61969/jai.1337500.
  32. E. Struble, M. Leon, E. Skordilis, “Intelligent Prevention of DDoS Attacks using Reinforcement Learning and Smart Contracts,” The International FLAIRS Conference Proceedings, 37(1), 2024, doi:10.32473/flairs.37.1.135349.
  33. X. Zhai, X. Chu, C. S. Chai, M. S. Y. Jong, A. Istenic, M. Spector, J.-B. Liu, J. Yuan, Y. Li, “A Review of Artificial Intelligence (AI) in Education from 2010 to 2020,” Complexity, 2021, 1–18, 2021, doi:10.1155/2021/8812542.
  34. C.-C. Lin, A. Y. Q. Huang, O. H. T. Lu, “Artificial intelligence in intelligent tutoring systems toward sustainable education: a systematic review,” Smart Learning Environments, 10(1), 2023, doi:10.1186/s40561-023-00260-y.
  35. W. Holmes, K. Porayska-Pomsta, K. Holstein, E. Sutherland, T. Baker, S. B. Shum, O. C. Santos, M. T. Rodrigo, M. Cukurova, I. I. Bittencourt, K. R. Koedinger, “Ethics of AI in Education: Towards a Community-Wide Framework,” International Journal of Artificial Intelligence in Education, 32(3), 504–526, 2021, doi:10.1007/s40593-021-00239-1.
  36. K. Zhang, A. B. Aslan, “AI technologies for education: Recent research & future directions,” Computers and Education: Artificial Intelligence, 2, 100025, 2021, doi:10.1016/j.caeai.2021.100025.
  37. L. Chen, P. Chen, Z. Lin, “Artificial Intelligence in Education: A Review,” IEEE Access, 8, 75264–75278, 2020, doi:10.1109/ACCESS.2020.2988510.
  38. G. N´apoles, I. Grau, R. Bello, M. Le´on, K. Vanhoof, E. Papageorgiou, “A Computational Tool for Simulation and Learning of Fuzzy Cognitive Maps,” in 2015 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), 1–8, IEEE, 2015, doi:10.1109/FUZZ-IEEE.2015.7338005.
  39. G. N´apoles, J. L. Salmeron, W. Froelich, R. Falcon, M. Leon, F. Vanhoenshoven, R. Bello, K. Vanhoof, Fuzzy Cognitive Modeling: Theoretical and Practical Considerations, 77–87, Springer Singapore, 2019, doi:10.1007/978- 981-13-8311-3 7.
  40. M. Le´on, “The Escalating AI’s Energy Demands and the Imperative Need for Sustainable Solutions,” WSEAS Transactions on Systems, 23, 444–457, 2024, doi:10.37394/23202.2024.23.46.
  41. G. N´apoles, F. Hoitsma, A. Knoben, A. Jastrzebska, M. Leon, “Prolog-based agnostic explanation module for structured pattern classification,” Information Sciences, 622, 1196–1227, 2023, doi:10.1016/j.ins.2022.12.012.

Citations by Dimensions

Citations by PlumX

Google Scholar

Crossref Citations

No. of Downloads Per Month
No. of Downloads Per Country

No related articles were found.

Journal Menu

Journal Browser


Special Issues

Special Issue on Digital Frontiers of Entrepreneurship: Integrating AI, Gender Equity, and Sustainable Futures
Guest Editors: Dr. Muhammad Nawaz Tunio, Dr. Aamir Rashid, Dr. Imamuddin Khoso
Deadline: 30 May 2026

Special Issue on Sustainable Technologies for a Resilient Future
Guest Editors: Dr. Debasis Mitra, Dr. Sourav Chattaraj, Dr. Addisu Assefa
Deadline: 30 April 2026