{
    "file_item": {
        "filepath": "security-guidance",
        "filename": "CERT-EU_Good_Practice-GP25-003.pdf"
    },
    "title": "Generative AI in Cybersecurity: Balancing Innovation and Risk",
    "publish_date": "03-07-2025 20:20:00",
    "description": "The integration of generative AI into cybersecurity operations represents both unprecedented opportunity and emerging risk. While these technologies offer powerful capabilities for threat analysis, incident response, and security automation, they simultaneously introduce new attack vectors that adversaries are rapidly exploiting. The cybersecurity community faces a critical challenge: how to harness AI's defensive potential whilst maintaining robust protection against AI-enabled threats. Organisations across sectors are grappling with questions of responsible adoption, governance frameworks, and risk management strategies. In this guidance, CERT-EU examines the dual nature of generative AI in cybersecurity, providing practical recommendations for security teams seeking to balance innovation with vigilance. Drawing from operational experience and threat landscape analysis, this resource offers actionable insights for navigating the complex decision-making process surrounding AI adoption in security contexts.",
    "url_title": "generative-ai-in-cybersecurity-balancing-innovation-and-risk",
    "content_markdown": "_**Some portions of this document were crafted with a little help from our internally deployed AI models (powered by\nLlama 3.3 - 70B). While we have fine-tuned and polished the content, this also showcases how Generative AI can be a\npowerful tool -- especially when running on your own infrastructure for greater control, security and privacy.**_\n\n## Generative AI\n\nGenerative AI refers to a class of artificial intelligence models designed to simulate human-like creativity and\nadaptability by generating new content, data, or outputs based on learned patterns from vast datasets. These AI systems\ncan be applied across a wide range of domains, from natural language processing to computer vision. For example,\nGenerative AI can create realistic images, draft human-like text, compose music, or design novel chemical compounds.\nNotable examples include large language models (LLMs) such as OpenAI's GPT series, Mistral Le Chat, Google's Gemini, and Meta's LLaMA, as\nwell as text-to-image generation tools like OpenAI's DALL-E and Stability AI's Stable Diffusion.\n\nLarge language models (LLMs) are advanced deep learning models that use self-attention mechanisms and multi-layered\narchitectures to understand and generate text. These models excel in tasks such as language translation, summarisation,\nquestion-answering, and creative content generation by analysing vast quantities of data and identifying complex\npatterns. Their strengths lie in their ability to capture nuanced contextual information, generate coherent and relevant\nresponses, and adapt to diverse domains.\n\nHowever, LLMs also have notable limitations. They are highly data- and computationally intensive, requiring substantial\nresources for training and fine-tuning. Additionally, they may produce plausible-sounding yet incorrect or nonsensical\nanswers (often referred to as hallucinations) and can be sensitive to input phrasing, leading to inconsistent results.\nFinally, they may inadvertently generate biased or harmful content due to biases present in their training data.\n\nSimilarly, text-to-image generation models utilise deep learning techniques to create visually coherent images based on\ntextual input. The strengths of these models include their ability to generate diverse and creative images, as well as\ncontributing to data augmentation and visual storytelling. However, like LLMs, their weaknesses include a dependence on\nlarge, well-annotated datasets for training, high computational requirements, and the potential to generate unrealistic\nor low-quality images. Furthermore, these models may struggle to accurately capture complex and abstract concepts\ndescribed in the textual input and, similar to transformers, may inadvertently propagate biases present in the training\ndata.\n\n## How does it work?\n\nGenerative AI might seem like magic at first, but it's actually the result of significant progress in deep learning. This progress is driven by rapid increases in computing power, access to large datasets, and better training techniques such as reinforcement learning and self-supervised learning. These advances have come together to make Generative AI a reality, rather than just an idea.\n\n### Text generation models\n\nAt the core of large language models are neural networks with millions or even billions of parameters (i.e., _model\nsize_), which are trained on vast amounts of text data. These _parameters_ define the connections between the nodes in\nthe network. Before an LLM can be used, it must undergo a training process, during which it is presented with massive\ndatasets (known as training sets) that allow the model to learn patterns, relationships, and structures within the\nlanguage. Through this, the model adjusts its parameters to minimise errors in prediction. Modern LLMs are trained using\nunsupervised learning and self-supervised learning, where the model learns by predicting the next word (or more precisely: _token_) in a\nsequence based on the prior context. LLMs can be broadly categorised into:\n\n- **Foundational models**\n\n  Foundational models, such as OpenAI's GPT series, Anthropic's Claude, Meta's LLaMA, and others -- serve as the base for a wide range of\n  applications. These models are pre-trained on enormous, diverse datasets sourced from books, articles, websites, and\n  other publicly available texts. Rather than learning specific tasks, foundational models learn a probabilistic\n  distribution of language -- in other words, they grasp how words, phrases, and concepts typically relate to one\n  another across a wide array of contexts. They are general-purpose models capable of understanding and generating text across\n  numerous domains without being tailored to specific tasks. Foundational models are highly versatile and can be applied\n  to various tasks, including translation, summarisation, creative writing, and even code generation.\n\n- **Fine-tuned models**\n\n  Fine-tuned models build upon foundational models but are adapted for specific tasks or domains. They undergo\n  additional training, or fine-tuning, using task-specific data or techniques such as Reinforcement Learning from Human\n  Feedback (RLHF). Fine-tuning enhances the model's performance in particular applications, allowing it to better meet\n  user needs. For instance, OpenAI's ChatGPT (based on the GPT-4 model) is a fine-tuned version of a foundational model,\n  optimised for conversational AI. It has been trained not only to generate text but also to handle dialogue, ensuring\n  that responses are contextually relevant, informative, and aligned with user intent. This fine-tuning process helps\n  reduce errors and improves the model's reliability in real-world applications, such as customer service, virtual\n  assistants, and personalised content generation.\n\nOnce training is complete, the model enters the _inference_ phase, where it generates predictions or completes tasks\nbased on new input data. During this stage, the model leverages its internal knowledge of language patterns and\nrelationships, acquired during training, to produce relevant and coherent output. For foundational models, this means\ngenerating responses based on a broad understanding of language and context, while fine-tuned models use additional\ntask-specific training to generate more tailored and accurate responses.\n\n### Visual generation models\n\nGenerative AI models are used to create visual content, such as images and videos, from textual descriptions. These models use neural networks to produce high-quality visual outputs. They are changing the creative industries by allowing dynamic content to be generated, whether it's a single image or a video sequence.\n\nThere are different types of models used for image and video generation. For example, Generative Adversarial Networks (GANs) and Diffusion Models, like DALL-E 3 and Stable Diffusion, are commonly used. Video models are more complex because they need to maintain both spatial and temporal coherence.\n\nImage generation models focus on converting text prompts into static images. Diffusion Models have become popular in recent years and work by refining random noise into structured images using text prompts. To train these models, large datasets of paired text and image data are needed. This allows the model to learn relationships between textual descriptions and their corresponding visual representations.\n\nVideo generation builds on the foundations of image generation but introduces the challenge of spatio-temporal relationships. This means ensuring that individual frames and transitions between them are coherent over time. Video generation models often adapt Diffusion Models to handle sequential frame generation while considering temporal flow. They need to learn to generate high-quality images for each frame and ensure movements, lighting, and objects remain consistent throughout the sequence.\n\nBoth image and video generation models rely on the concept of latent space during inference. Latent space is an abstract representation of the learned relationships between visual and textual elements. In image generation, the model samples from this latent space to produce a single visual output. For video generation, the latent space also encodes temporal dynamics, enabling the model to generate a sequence of frames that align with the input text while ensuring smooth transitions.\n\nTo achieve this, video generation models are trained on vast video-text paired datasets. The focus is on learning to generate not just realistic images but also seamless motion and narrative progression. This results in dynamic content that feels natural while being driven by the input text. By understanding how these models work, you can explore their potential applications in various industries and creative projects.\n\n## Future outlook\n\nArtificial intelligence is advancing rapidly, with transformative innovations reshaping industries and unlocking new\nopportunities. The growth of open-source and open-weights models has significantly expanded AI's accessibility and\napplications. Unlike proprietary systems, these frameworks allow organisations, researchers, and developers to deploy,\nadapt, and refine AI tools without restrictive licensing, promoting greater autonomy. This shift is particularly evident\nin Generative AI, where advanced open source projects such as DeepSeek are driving significant breakthroughs. These\nmodels now enable diverse applications, from language processing to image and video generation, empowering businesses\nand\nindividuals to innovate cost-effectively.\n\nCost-effective AI models are now rivaling proprietary models in performance. Research into self-supervised,\nunsupervised, and reinforcement learning is advancing rapidly, while breakthroughs in multimodal AI -- integrating text,\nimages, audio, video, and even interactive environments -- are pushing the boundaries of creativity and problem-solving.\n\nAt CERT-EU, we are pursuing in-house AI projects to optimise operations and deepen our expertise in these technologies.\nThis hands-on approach ensures that our advice on AI systems is both informed and actionable. While AI's democratisation\noffers opportunities to enhance innovation, personalise experiences, and automate tasks, it also presents significant\nrisks, such as the spread of disinformation, misuse by malicious actors, and ethical dilemmas surrounding synthetic\ncontent. As the technology evolves, it is crucial to address these challenges proactively to ensure its responsible\ndevelopment.\n\n\n\n\n\n\n\n\nGenerative AI holds immense transformative potential, reshaping various sectors, including cybersecurity. This document\nspecifically focuses on its implications for Union entities, examining how Generative AI is revolutionising both the\ndefensive and offensive aspects of cybersecurity. By exploring how these technologies can strengthen protective measures\nwhile also enabling new threats, we aim to assess their impact on organisations within the Union.\n\nOur goal is to propose actionable recommendations that will help direct and coordinate the efforts of Union entities in\neffectively harnessing the benefits of Generative AI, while mitigating its associated risks. Given the rapid pace of\ninnovation, these insights reflect the landscape as of mid-2025, with the understanding that ongoing developments may\nsubstantially alter the threat environment.\n\n## Defensive applications of AI\n\nArtificial intelligence offers powerful tools to counter sophisticated cyber threats, mitigating the traditional\n\u201cDefender's Dilemma\u201d[^1]\n-- where attackers historically retain the upper hand. AI systems excel at transforming vast datasets\ninto actionable intelligence, bolstering capabilities such as malware detection, vulnerability identification, and\nthreat analysis. By automating routine tasks and accelerating response times, these technologies enable security teams\nto operate with heightened efficiency and precision.\n\nGenerative AI further augments defences. For example, it can simulate realistic cyber-attack scenarios, such as phishing\ncampaigns or ransomware simulations, to rigorously test and train personnel. Beyond training, generative models\nfacilitate the creation of adaptive honeypots that mislead attackers while gathering tactical intelligence. Large\nlanguage models add value by detecting subtle patterns in data, such as log file anomalies, empowering analysts to\nprioritise risks and uncover hidden correlations.\n\n## Offensive applications of AI\n\nThe same technologies empowering defenders are increasingly weaponised by adversaries. Generative AI, for instance, is\nexploited to craft hyper-realistic social engineering campaigns, including personalised phishing emails, SMS scams, and\ndeepfake audio/video. Beyond deception, these models automate the discovery of software vulnerabilities -- even\nuncovering novel attack vectors, and generate functional code for malware or evasion techniques[^2]. Tools like WormGPT or FraudGPT,\nfor example, illustrate how readily AI can be adapted to scale malicious activities.\n\nThe escalation of AI-driven disinformation and cyberattacks presents acute challenges for the Union entities. This\nevolving threat landscape underscores the urgent need for proactive, AI-enhanced security strategies, such as\nbehaviour-based anomaly detection and predictive threat modelling, to counter adversarial innovation.\n\nIt is essential for CERT-EU to communicate not only the risks posed by Generative AI but also the opportunities it\noffers to enhance the resilience and cybersecurity capabilities of Union entities. Collaboration among these entities\nwill be crucial for sharing best practices, investing in development, and establishing ethical frameworks for AI use. By\nprioritising human-AI collaboration and maintaining rigorous oversight, CERT-EU and Union entities can set a benchmark,\nfostering a secure and innovative digital ecosystem within the public institutions of the European Union.\n\n\n\n\n\n\n\nAs AI continues to evolve, organisations are now focused on integrating these technologies into their existing\ncybersecurity teams and systems. The challenge lies in ensuring the effective adoption of AI solutions, aligning them\nwith strategic objectives, and optimising overall security efforts.\n\nA 2024 survey by the Cloud Security Alliance (CSA)[^3] found that 63% of\nsecurity professionals believe AI will enhance security measures. However, adoption remains in its early stages, with\nonly 22% of organisations currently using Generative AI, though 55% plan to implement it within the next year.\n\n## AI's impact on security operations\n\nAI-driven automation is reshaping security operations by improving threat detection, accelerating incident response, and\nreducing operational burdens. While concerns about AI replacing human roles exist, most professionals see it as a\ncomplementary tool. Only 12% believe AI will fully replace their roles, whereas 30% expect it to enhance their skills,\nand 28% see AI as broadly supporting their work. However, 51% warn against over-reliance, emphasising the need for human\noversight in AI-driven security strategies.\n\nOrganisations are deploying AI to address workforce shortages and improve security efficiency. Around 36% of security\nteams use AI to bridge skill gaps, while 26% prioritise faster threat detection. Other key objectives include improving\nproductivity and reducing misconfigurations. Generative AI is increasingly used for automated rule creation, attack\nsimulations, and compliance monitoring, but its adoption introduces risks such as data manipulation and adversarial\nattacks. Effective governance frameworks are necessary to ensure responsible AI deployment.\n\n## Challenges in AI adoption\n\nDespite its advantages, AI adoption in cybersecurity presents significant challenges. A primary concern is the shortage\nof skilled professionals who can effectively manage and secure AI systems, with 33% of organisations citing skill gaps\nas a major barrier. The complexity of AI models requires specialised expertise to train, maintain, and interpret their\noutputs. Additionally, 38% of security professionals highlight data quality issues, including unintended bias in AI\nmodels, while 25% cite privacy risks as a growing concern.\n\nAI security risks extend beyond implementation challenges. Security professionals are increasingly aware of\nvulnerabilities such as data poisoning, adversarial attacks, and AI-generated misinformation. Approximately 28% of\nrespondents express concerns about data poisoning attacks, which could manipulate AI models to generate false outputs.\nRegulatory and compliance considerations further complicate AI deployment, requiring organisations to navigate evolving\nsecurity frameworks.\n\n## Future trends in AI-Driven cybersecurity\n\nAI adoption is expected to accelerate, with organisations increasingly exploring Generative AI for cybersecurity\napplications. Automated rule creation, attack simulations, and compliance monitoring are among the most common use\ncases. While AI-driven automation will continue to expand, security teams must balance efficiency gains with oversight\nmechanisms to mitigate potential risks.\n\nGovernance structures are evolving to support responsible AI deployment. Many organisations are establishing dedicated\nteams to oversee AI implementations, ensuring compliance with security policies and regulatory requirements. Executive\nleadership plays a crucial role in AI adoption, with 82% of organisations reporting strong leadership support for AI\ninitiatives. However, a gap remains between executive-level strategy and operational execution, reinforcing the need for\nclear guidelines and AI-specific training programmes.\n\nAs AI adoption progresses, cybersecurity strategies must adapt to emerging threats. The increasing sophistication of\nAI-driven cyberattacks necessitates continuous advancements in AI-based defence mechanisms. Security teams must also\naddress AI security risks, improve transparency in AI decision-making, and implement safeguards against adversarial\nmanipulation. The widespread integration of AI into cybersecurity marks a transformative shift, requiring a strategic\napproach to maximise its benefits while mitigating associated risks.\n\n## European Union initiatives and investments in AI\n\nThe European Commission has made significant strides in promoting the secure and compliant use of Generative AI\ntechnologies. A notable example is its approval of the Azure OpenAI service for handling Sensitive Non-Classified Data,\nwhich ensures that user prompts remain within the EU and are excluded from model training. Hosted in regional cloud\nenvironments under Cloud Broker Contracts, this service strikes a balance by providing robust privacy safeguards without\nthe complexities and resource-intensive demands of managing local models.\n\nSimilarly, the Commission has\nimplemented GPT@EC[^4],\ndeploying AI models within a secure, local infrastructure to enhance\noperational efficiency while adhering to EU data protection standards. Complementing these efforts, CERT-EU has\nintegrated internal large language models to power AI-driven projects, ensuring sensitive data remains protected while\nboosting productivity and supporting internal initiatives to reduce manual workloads and enhance decision-making\nprocesses\n\nLooking ahead, the European Union has pledged \u20ac200 billion in AI\ninvestments[^5], aiming to build critical\ninfrastructure such as AI training gigafactories within the region. This ambitious plan highlights Europe's commitment\nto advancing AI innovation while maintaining a strong focus on data privacy and regulatory compliance.\n\n\n\n\n\n\n\n\n\nAs this document focuses on the impact of Generative AI on cybersecurity and its relevance to Union entities, several\nupdated use cases in cybersecurity are highlighted, with an emphasis on ongoing AI-driven initiatives.\n\n## Improving threat detection\n\nAI systems enhance cybersecurity by refining detection algorithms and generating new rules based on the latest threat\nintelligence. By leveraging AI-driven insights, security teams can improve detection capabilities, strengthening\ndefences against cyberattacks.\n\nLog analysis is a critical aspect of a cybersecurity analyst's role, and large language models (LLMs) can significantly\nstreamline this process. AI can sift through vast amounts of log data, identifying anomalies, outliers, and correlations\nthat may indicate security breaches. Automating log analysis reduces the time and effort required, allowing analysts to\nfocus on high-priority threats rather than manually searching for patterns. Additionally, AI systems can detect\nconnections between seemingly unrelated events, providing a comprehensive view of the security landscape and enabling\nmore effective threat responses.\n\nAI also holds promise in enhancing detection rules by analysing large datasets, such as system logs, to uncover patterns\nthat may be missed by human analysts. However, false positives remain a major concern. Even with a low false positive\nrate of 0.1%, the sheer volume of log entries -- often in the hundreds of millions -- can result in an overwhelming number of\nalerts, necessitating careful tuning of AI models.\n\nCERT-EU is exploring the use of retrieval-augmented generation (RAG)-based systems to assist analysts in creating and\nrefining detection rules. These systems leverage existing rules to generate new ones, which analysts can review and\nimplement. Other AI-driven platforms already being used for threat detection and response include Microsoft Security\nCopilot[^6]\nand\nPalo Alto Networks' Cortex\nXSOAR[^7],\ndemonstrating AI's growing role in strengthening cybersecurity operations.\n\n## Supporting analysis\n\nLLMs have also enabled analysts to work more efficiently and accurately in various aspects of their jobs. One key area\nwhere they have demonstrated their potential is in the deobfuscation of malicious code. Attackers often obfuscate their\ncode to evade detection, but LLMs can assist analysts in identifying patterns and decoding hidden algorithms, providing\nvaluable insights into the attacker's intent and revealing the true nature of the threat, thereby significantly speeding\nup the investigation.\n\nAdditionally, LLM transformers show a remarkable ability to correlate data from various sources and fields as they have\nbeen trained on diverse datasets. This vast training corpus enables the AI model to extract and synthesise information\nfrom various seemingly unrelated sources. By leveraging its deep learning capabilities, LLMs can then identify\nconnections, patterns, and insights across these different sources. As a result, the AI model has proven to be a\nvaluable tool in solving problems, providing novel insights, and identifying correlations that would otherwise be easy\nto miss.\n\nCode analysis and reverse engineering are also areas where LLMs can provide substantial assistance. With their extensive\nknowledge base, LLMs can evaluate software code and explain its operation. In the case of reverse engineering, LLMs can\nhelp dissect complex obfuscated code and provide insights into its functionality and purpose. By understanding how a\npiece of malware or exploit operates, analysts can develop effective countermeasures to protect their systems and\nnetworks, while also saving time during the investigation.\n\nFor example, VirusTotal's[^8] integration\nof\nthe Sec-PaLM model now generates natural language summaries of code snippets, enabling analysts to identify potential\nvulnerabilities more quickly. Similar features are likely to be integrated into many reverse engineering, sandboxes, and\nanalysis tools in the future.\n\n## Automating threat intelligence\n\nLLM transformers can greatly enhance the process of generating threat intelligence reports by automating the collection,\nanalysis, and summarisation of relevant data. This not only saves time and effort, but it also ensures that the\ninformation presented to cybersecurity teams is accurate, up-to-date, and easily digestible. Armed with this\nintelligence, defenders can make more informed decisions and take proactive measures to protect their systems.\n\n## Coding and documentation\n\nLLM transformers have already shown their impact on the field of software development. These AI models assist developers\nwith code completion, bug detection, security vulnerability identification, and automatic documentation. By suggesting\nsecure code snippets and identifying flaws early, LLMs help accelerate the development process and improve security.\n\nTools like GitHub Copilot X[^9] or Amazon\nCodeWhisperer[^11] are enhancing developer productivity and integrating security\nbest\npractices into workflows. While not replacing human developers, these tools provide significant support, especially in\npreventing common vulnerabilities such as SQL injection or cross-site scripting.\n\nGenerative AI also automates routine tasks like writing documentation and generating unit tests, enabling developers to\nfocus on more complex security issues. However, since LLM-generated code is based on public datasets, there is a risk of\ninheriting bugs or vulnerabilities, making human oversight essential to ensure secure, production-ready code.\n\nFor example, CERT-EU has deployed internal copilot tools that leverage the use of internal AI models to assist\ndevelopers in writing secure code and identifying potential vulnerabilities. These tools have significantly improved the\nquality of code produced and reduced the time spent on manual code reviews.\n\n## Enhancing cybersecurity training\n\nThe sophisticated natural language processing capabilities of large language models are playing an increasingly\nimportant role in upskilling cybersecurity personnel. AI-powered systems now offer contextualised explanations of\ncomplex cybersecurity concepts, enabling junior staff to bridge knowledge gaps more quickly. For example, AI platforms\ncan provide personalised guidance on identifying and responding to emerging threats. By offering real-time support,\nthese models help less experienced team members contribute more effectively to threat analysis and mitigation, thereby\nenhancing the overall efficiency of security teams.\n\nMoreover, many institutions are now offering more frequent internal training sessions to help staff better understand\nand properly utilise AI tools. These training programmes are designed to ensure that teams can leverage AI tools\neffectively, enhancing their ability to respond to cybersecurity challenges in a rapidly evolving landscape.\n\n## Content generation\n\nFinally, one of the most evident applications of Generative AI is the creation of high-quality content across a range of\ndomains, including the automatic generation of technical documentation, corporate communications, and presentations. The\nability of AI-driven content generation platforms to understand context and produce human-like text can enhance\norganisations\u2019 approaches to content creation. The anticipated widespread adoption of Generative AI is largely due to\nits impressive capacity to save time, reduce costs, and increase overall efficiency in producing a variety of content\ntypes.\n\nIn the field of cybersecurity, Generative AI can be employed to draft post-incident analyses, summarise threat\nintelligence feeds, and produce tailored security advisories. It also supports the automation of incident reports,\nenabling security teams to focus on more strategic tasks. By streamlining the drafting process, organisations can ensure\nthat critical information is communicated clearly and promptly, facilitating quicker and more effective responses to\nemerging threats.\n\nSimilarly, AI-driven content generation can enhance the quality and efficiency of presentations. By leveraging data,\nGenerative AI can dynamically generate visuals, suggest relevant talking points, and recommend persuasive storytelling\ntechniques to engage audiences. This not only simplifies the process of creating presentations but also increases their\noverall impact and effectiveness.\n\nSimilarly to code generation examples, there are already several products either available or being rolled out that\npropose such AI-based enhancements and solutions. These include, for instance, Microsoft 365\nCopilot[^12] and\nGoogle AI-powered Workspace Features[^13].\n\n\n\n\n\n\n\n\n\nGenerative AI models usually require significant computational resources to function effectively. Choosing the right\ninfrastructure is crucial to balancing performance, scalability, cost, and security. There are several deployment\noptions available, each with distinct advantages and trade-offs.\n\n## Public closed-source models \u2013 paid or free\n\nThe AI industry has seen a surge in public closed-source models offered by major tech companies. These models provide\npowerful AI capabilities but often function as black-box solutions, limiting user control over data processing and\nstorage.\n\nThe dominant market model today revolves around \"free\" closed-source AI services, such as ChatGPT, DALL-E, Midjourney,\nand Google Gemini. These platforms make AI highly accessible but come with significant data privacy concerns. Their\nterms of use often indicate that input and output data may be stored outside the EU and could be used for further\ntraining and fine-tuning of the models. As a result, organisations handling sensitive information must assume that any\ndata provided through these services could become public knowledge.\n\nAdditionally, the emergence of cost-effective AI models from competitors like China's DeepSeek has intensified market\ncompetition, prompting tech giants to reassess their offerings and pricing structures. While these services offer ease\nof use and accessibility, organisations must carefully evaluate whether the trade-offs in data sovereignty and\ncompliance align with their operational and regulatory needs.\n\n## Locally-hosted open-source models\n\nDeploying open-source AI models on local infrastructure has become a key strategy for organisations prioritising data\ncontrol, security, and customisation. Models like LLaMA 3 and Mistral can be hosted on-premises or in private clouds,\nensuring compliance with data sovereignty regulations while avoiding reliance on external providers.\n\nHowever, maintaining these systems requires high-performance GPUs and skilled personnel, leading to the growing adoption\nof AI colocation services. These services provide access to cutting-edge computing infrastructure without the overhead\nof in-house data centre management.\n\nAs AI advances, organisations are increasingly balancing performance, cost, and security by adopting self-hosted\nopen-source models and colocation solutions, ensuring greater control over AI applications while maintaining\nscalability.\n\n## Privacy-focused commercial closed-source models with specific conditions of use\n\nIn response to increasing concerns over data privacy and regulatory compliance, several tech companies have introduced\nprivacy-focused AI services that adhere to strict data handling policies. These cloud providers now offer solutions that\nensure user data remains within specified regions and is not used for training or fine-tuning models, striking a balance\nbetween leveraging advanced AI capabilities and maintaining control over sensitive data.\n\nFor organisations handling sensitive information, these services offer an opportunity to benefit from AI technologies\nwhile ensuring privacy compliance. Additionally, privacy-focused commercial models with negotiated terms of use are\ngaining traction. These models come with specific configurations and agreements that differ from public closed-source\nmodels, providing stronger data protection but imposing more stringent conditions for non-compliance. As these offerings\nbecome more prevalent, organisations must carefully assess the terms of service to ensure they align with privacy\nrequirements and legal obligations within the evolving regulatory landscape.\n\nWhen selecting a deployment option for Generative AI models, organisations must carefully evaluate their specific needs\nand privacy requirements. One promising option for Union entities is the adoption of privacy-focused commercial models with\ncustomised terms of use. These models offer stronger data protection than public closed-source solutions, but with\nstricter conditions that may carry serious consequences for non-compliance.\n\n\n\n\n\n\n\n\nRisks associated with the use of Generative AI can broadly be divided into two main categories:\n\n- Risks related to its use within an organisation.\n- Risks resulting from its use by others, including malicious actors.\n\n## Risks of using Generative AI\n\nSpecific risks may arise from the potential use of Generative AI technology by the staff of Union entities. As with its\nbenefits, the emphasis remains firmly on cybersecurity, with a particular focus on these organisations.\n\n### Indirect prompt-injection attacks\n\nAs Generative AI technology evolves, new possibilities and risks emerge. The recent development of various plugins and\ninterfaces for external data sources that\ncan be used in conjunction with some of the large language models -- or even independent AI Agents -- increase their\ncapabilities but also introduce new\nrisks.\n\nOne of these risks is the possibility of indirect prompt-injection attacks. When\nmodels are able to use external data -- websites, documents, emails, etc., such external data may potentialy be under\nthe control of\nmalicious actors. This can allow an attacker to attempt to influence the model's output by carefully crafting their\ninput\nor \"prompt\", often embedding hidden instructions or biases. The AI model then inadvertently generates output that could\npotentially spread misinformation, reveal sensitive information, or produce other undesirable outcomes. Despite the\ninput appearing harmless or neutral to a human observer, it can result in manipulated outputs, thus presenting\nsignificant security concerns in the application of AI technologies.\n\nIndirect prompt-injection attacks are already occurring in the wild, with various tools and setups allowing LLMs to\naccess\nexternal data sources being used as vectors for these attacks. Examples include hidden text in web pages, inside\ndocuments or emails that are then provided as input to LLMs by unsuspecting\nusers.\n\nA significant challenge is that the existing defences are not currently equipped to effectively counter these attacks.\nThe\nsubtlety of the manipulation makes detection extremely difficult, especially as the injected prompts often appear\nharmless or neutral to human observers or are not easily visible at all. While it is possible to configure the models to\nignore certain types of these attacks or specific prompts, there is no obvious way to create a permanent fix. Users\nshould be cautious when using AI tools on any input that may have been subject to malicious modification (e.g., web\npages, external documents, incoming emails, etc.).\n\n### Disclosure of sensitive data\n\nThe use of freely available, closed-source AI language models, such as ChatGPT, poses potential risks to sensitive data\nsubmitted in user prompts. As users interact with the model, they may inadvertently input confidential or personally\nidentifiable information (PII) while seeking assistance or answers. Since this information is typically stored to enable\nthe model to process and generate responses, there is a risk that sensitive data could be exposed, either through data\nbreaches or during the training of future versions of the AI models. Without proper data anonymisation and privacy\nsafeguards in place, such information could be misused by unauthorised parties, leading to identity theft, financial\nfraud, or reputational damage for both individuals and organisations involved.\n\nFor instance, OpenAI\u2019s current terms of use[^14] specify that while OpenAI does\nnot use API content to improve their services, they may use non-API content (i.e., prompts and outputs from ChatGPT) for\nthis purpose. Therefore, if confidential or sensitive data is entered as part of a ChatGPT prompt, it could eventually\nbe exposed. OpenAI states that requests submitted via their API will be stored for 30\ndays[^15] and not used for training.\nHowever, there is no guaranteed proof of compliance or transparency regarding OpenAI\u2019s future plans.\n\nIn the event of a cyberattack on the infrastructure of an AI language model, there is a significant risk of data\nleakage. Such a breach could expose sensitive and private user information, including personal details, confidential\nconversations, and intellectual property. The consequences of such exposure could be wide-ranging, including compromised\nprivacy, loss of user trust, and potential legal ramifications.\n\nOrganisations must also be vigilant in how AI tools are deployed within their environments. Employees should be trained\nnot to enter sensitive data into public models, and technical controls -- such as data masking, secure API gateways, and\naudit logging -- should be implemented where possible.\n\nThe European Commission\u2019s AI Act[^16] seeks to\nenforce strict standards around AI deployment and data protection. By complying with its provisions, organisations can\nreduce the risk of unauthorised access, enhance system transparency, and demonstrate accountability in how sensitive\ndata is handled. This regulatory framework fosters public trust in AI technologies and encourages investment in\nprivacy-preserving solutions.\n\n### Copyright violations\n\nGenerative AI technologies, such as text and image generation models, have raised concerns about potential copyright\nviolations as they become increasingly adept at creating content that closely resembles human-authored works. In the\nrealm of text generation, AI-powered tools can produce articles, stories, or even poetry, often blurring the lines\nbetween human creativity and synthetic, machine-generated output. This raises questions about the originality of the\ncontent and whether the AI system has unintentionally reproduced or closely mimicked copyrighted materials.\n\nFor instance, if a text-generation AI model creates a story that closely resembles a popular novel, the copyright holder\nof the original novel may claim infringement, arguing that the AI-generated work could be perceived as a derivative of\ntheir copyrighted material.\n\nSimilarly, image-generation models have the capability to create visually appealing artwork, designs, and even\nphotorealistic images. These AI-generated images could infringe upon copyrighted visual content if they closely resemble\nexisting works, such as paintings, photographs, or graphic designs. For example, if an image-generation AI model were to\ncreate an artwork strikingly similar to a famous painting, it could lead to copyright disputes between the original\nartist and the creator of the AI-generated piece. Moreover, these concerns extend to the potential appropriation of\nelements from multiple copyrighted works to create a new image, which could lead to multiple copyright violation claims.\n\nIn both cases, the increasing sophistication of Generative AI technologies complicates the legal landscape surrounding\ncopyright protection, as it becomes more challenging to determine the true authorship and originality of content.\n\nAdditionally, in some instances, the models powering Generative AI technologies are known to be trained on copyrighted\ncontent without the explicit approval of the authors. This raises additional concerns, as the organisations behind these\nmodels could be held liable for potential copyright infringement. By using copyrighted material to train their AI\nsystems, organisations may inadvertently propagate the unauthorised reproduction or adaptation of protected works,\nopening themselves up to potential litigation. As a result, there is a growing need for more robust and transparent\ncontent acquisition policies to ensure that the data used to train AI models is either appropriately licensed or falls\nunder the scope of fair use.\n\n### False or inaccurate information\n\nAI language models have become increasingly adept at generating high-quality text. However, these models have flaws, and\nthe risk of providing false or inaccurate information remains\nsignificant[^17]. As AI language models are trained on vast\namounts of data from the internet, they are susceptible to absorbing and perpetuating the biases, misconceptions, and\ninaccuracies present in that data. It is also important not to confuse Natural Language Processing (NLP), which these\nmodels excel at, with Natural Language Understanding (NLU), a significant challenge in AI research. A system trained\nsolely on form has a priori no way to learn meaning[^18]. Consequently,\nusers of these models must be aware of the potential\npitfalls and exercise critical thinking when interpreting generated text.\n\nOne primary concern with AI language models is the possibility of bias. As these models learn from the data they are\ntrained on, any biases present in the training data are likely to be absorbed and perpetuated by the model. This could\nmanifest as gender, racial, or political biases, among others, and can lead to the generation of text that is offensive\nor perpetuates harmful stereotypes. In some cases, these biases may even cause the AI to provide misleading or outright\nfalse information, leading users astray and potentially reinforcing pre-existing biases.\n\nSimilarly, when generating computer code in various programming languages, Large Language Models (LLMs) often provide\ncode containing errors or security vulnerabilities. This is primarily due to the training data these models are exposed\nto, which may include a diverse array of programming languages, styles, and quality levels. As a result, LLM-generated\ncode may not always adhere to best practices or conform to the latest security standards. Additionally, these models\nlack the inherent ability to understand the context or specific requirements of a given task, which may lead to the\nproduction of code that is unsuitable, flawed, or even dangerous. Therefore, it is crucial for developers to carefully\nreview and validate any code generated by LLMs before incorporating it into their projects to mitigate potential risks\nand ensure the safety and integrity of their software applications.\n\nAnother concern is the phenomenon of \"hallucinations\" where AI language models generate text that appears plausible but\nis entirely fabricated or lacks a factual basis. These hallucinations can occur for various reasons, such as the model\ntrying to fill in gaps in its knowledge or attempting to provide a coherent response to an ambiguous or unfamiliar\nprompt. While these hallucinations can sometimes be relatively harmless, in other instances, they can lead to the\ndissemination of false information or contribute to the spread of misinformation.\n\n### Hype abuse\n\nThe rapid advancements in Generative AI technology and the surrounding hype have led to a surge in public interest and\nadoption. While these innovations undoubtedly offer numerous benefits and transformative potential, the excitement can\nalso lead to possible pitfalls. With increased hype, bad actors may exploit the situation by creating fake applications\nor investment schemes, capitalising on the popularity of recognisable AI brand names to deceive users and achieve\nmalicious objectives.\n\nOne such pitfall is the emergence of fake ChatGPT apps on Android and iOS platforms. These counterfeit apps, disguised\nas popular AI language models, may carry out invasive data harvesting activities. Unsuspecting users who download and\ninteract with these malicious apps may inadvertently expose their personal information, including messages, contacts,\nand browsing history. The harvested data can then be used for various nefarious purposes, such as identity theft,\ntargeted advertising, or even extortion. This underscores the importance of exercising caution when downloading mobile\napplications and ensuring they originate from trusted sources and developers.\n\nAnother potential pitfall linked to the hype around Generative AI is the creation of fictitious cryptocurrency tokens\nusing recognisable AI brand names. Bad actors may design and market these tokens to lure in unsuspecting investors, who\nmay believe they are investing in a promising AI venture. Once the scammers accumulate a substantial amount of funds,\nthey may disappear, leaving the investors with worthless tokens and significant financial losses. This highlights the\nneed for investors to conduct thorough research and due diligence before committing to any investment, particularly in\nemerging technologies like AI and cryptocurrencies.\n\n### Over-relying on technology\n\nOver-relying on Generative AI technology presents several potential dangers that could significantly impact\norganisations. One key concern is the possible erosion of competence among staff. As AI systems become increasingly\ncapable of handling tasks traditionally carried out by humans, employees may become more reliant on these technologies.\nThis dependence could lead to a decline in critical thinking and problem-solving skills, making staff less versatile and\nadaptive when faced with new challenges. Furthermore, as AI takes over routine tasks, workers may lose the ability to\nperform these manually, resulting in a loss of valuable expertise.\n\nAnother issue is the overconfidence in the quality of output generated by AI. Due to inherent limitations in AI models,\nsuch as the token limits that restrict the amount of information a language model can \"remember\", the generated content\nmay not be as accurate, comprehensive, or contextually appropriate as users might expect. This could lead to situations\nwhere AI-generated content is accepted at face value, potentially causing misinformation or poorly informed decisions.\n\nThe over-reliance on AI technologies may also manifest as a failure to account for policy or political decisions that\ncould limit their use. Governments and regulatory bodies are increasingly scrutinising the implications of AI on\nprivacy, security, and social equality, as seen with regulations like the AI\nAct[^19]. Consequently, new policies or\nregulations may be introduced, imposing restrictions on the development, deployment, or use of AI technologies.\nOrganisations that become overly dependent on AI systems may find themselves ill-prepared to adapt to these changes,\nleading to potential operational disruptions.\n\nFinally, as highlighted in the benefits section, when using LLM tools for programming, it is essential to remember that\nthe generated code may contain bugs or be insecure or unsuitable. Extra care must be taken when allowing staff and\ncontractors to use LLMs for developing applications. The emphasis should be on ensuring thorough validation and testing\nof the generated code to mitigate potential risks.\n\n### LLMs opinions, advice, and moral values\n\nLarge Language Models (LLMs), such as ChatGPT or DeepSeek, should not be relied upon for opinions, advice, or moral\nguidance due to\nthe inherent limitations in their design and the nature of their training data. While LLMs are powerful AI tools, they\nare not human beings with emotions, life experiences, or ethical systems. Instead, they are complex algorithms built to\ngenerate humanlike text based on patterns and associations identified within vast amounts of data.\n\nOne of the primary reasons LLMs are unsuitable for providing opinions, advice, or moral guidance is that their responses\nare derived from the datasets used in their training. These datasets consist of vast amounts of text from a wide variety\nof sources, which may contain conflicting opinions, values, and perspectives. When an LLM encounters such\ncontradictions, it may struggle to produce a coherent and consistent response. As a result, the output may seem random,\nas the model attempts to balance opposing viewpoints or may simply reproduce popular opinions without understanding the\nunderlying reasons or nuances.\n\nFurthermore, LLMs are incapable of forming independent opinions or moral judgments. They do not have the capacity to\ncritically analyse complex issues or empathise with human emotions, both of which are essential when providing sound\nadvice or ethical guidance. Relying on an LLM for such matters could lead to misguided or superficial conclusions that\nfail to address the unique complexities of a given situation.\n\nIt is therefore not surprising that, for example, Deepseek has imposed strict guardrails on the values that LLMs must\nreflect, aligning them with those of the Chinese Communist Party, as part of its broader efforts to enforce censorship\nand control over AI-generated content[^20].\nAfter all, an LLM will inevitably reflect the moral values embedded in its training data and shaped by human feedback\n(Reinforcement Learning with Human Feedback -- RLHF). Consequently, for any generated text intended for political\npurposes, it may be wise to verify whether it aligns with the general vision, policy, and strategy of Union entities.\n\n## Risks from adversarial use of Generative AI technology\n\nSpecific risks arise from the use of Generative AI technology by malicious actors. As previously mentioned, the focus\nremains on cybersecurity, particularly with regard to Union entities.\n\n### Privacy issues\n\nPersonally identifiable information (PII) can inadvertently be included in the training datasets of generative AI models\nwhen data is collected from a broad range of sources, such as websites, forums, social media, and other digital\nplatforms. This data may not always be properly anonymised or sanitised before being used to train the models. As a\nresult, PII \u2014 including names, addresses, telephone numbers, email addresses, or other sensitive details \u2014 could become\nembedded within the model\u2019s training data, potentially allowing it to be traced back to specific individuals.\n\nWhen these models are deployed, there is a risk that PII could be unintentionally disclosed through generated outputs.\nHowever, malicious actors may also deliberately exploit generative AI models to extract or reconstruct sensitive\ninformation, using techniques such as prompt injection or model inversion. These attacks are designed to probe the model\nfor private or confidential data that may have been memorised during training.\n\nThis presents a twofold privacy risk: firstly, the unauthorised disclosure or targeted extraction of sensitive\ninformation could have serious consequences for the individuals concerned; secondly, the generated content may be\ninaccurate or misleading, resulting in misinformation or the misidentification of individuals. Both scenarios pose\nsignificant threats to privacy, trust, and the safe adoption of AI technologies \u2014 particularly in sensitive sectors such\nas government, healthcare, and finance.\n\n### More advanced cyberattacks\n\nGenerative AI technologies could also give rise to new methods for conducting cyberattacks. As AI systems become more\nsophisticated, they can be increasingly exploited by malicious actors to facilitate attacks and exploit vulnerabilities\nin various ways.\n\nOne such method involves using AI to generate phishing content. By harnessing natural language processing and generation\ncapabilities, cybercriminals can craft highly convincing emails, text messages, and social media posts that appear to\ncome from legitimate sources. These AI-generated messages can be specifically tailored to target individuals, increasing\nthe likelihood of them falling victim to the scam. Furthermore, AI can automate the process of sending phishing\nmessages, allowing attackers to target a larger number of potential victims.\n\nSocial engineering attacks can also be enhanced by AI-generated voice and video deepfakes. These realistic forgeries can\nbe used to impersonate executives, celebrities, or other influential figures, manipulating victims into providing\nsensitive information or taking actions that benefit the attacker. Deepfake technology can also be employed to create\nmore convincing phone scams or video calls, further increasing the likelihood of a successful attack.\n\nAdditionally, AI technologies can be used to improve malware, making it more difficult to detect and more effective in\nits operations. For example, AI algorithms can analyse existing malware and identify patterns likely to be flagged by\nantivirus software. Based on this analysis, AI can then generate new, stealthier malware variants that are harder to\ndetect and better exploit system vulnerabilities.\n\nAI can also facilitate cyberattacks through more efficient vulnerability detection and fuzzing. By using AI-powered\ntools, attackers can rapidly discover security weaknesses in software or network infrastructure, much faster than\ntraditional methods. This allows them to identify and exploit vulnerabilities before they are patched, increasing the\nlikelihood of a successful attack.\n\nFurthermore, AI can be used to automate and optimise password cracking. By employing machine learning algorithms,\nattackers can recognise patterns in password creation and generate more effective password dictionaries, significantly\nspeeding up the cracking process. This can drastically reduce the time it takes to gain unauthorised access to accounts,\nmaking it more difficult for security professionals to respond effectively.\n\nFinally, the development of freely available Generative AI tools has inadvertently lowered the entry barriers for new\nmalicious actors in the cybercrime ecosystem. With minimal technical expertise, individuals can exploit the capabilities\nof advanced AI models to conduct various illicit activities, such as generating phishing emails, creating realistic\ndeepfakes, or producing fake news. This democratisation of access to powerful AI-driven tools amplifies the reach and\nimpact of cybercrime, cyber espionage, and other malicious activities. It also poses significant challenges for\ncybersecurity professionals, law enforcement, and policymakers, as it allows a wider range of actors to participate in\nthese activities.\n\n### Disinformation\n\nThe powerful capabilities of Generative AI models come with significant risks when misused for disinformation campaigns.\nThese models can impersonate public figures and create highly convincing narratives, making them potent tools for\nspreading false and misleading information. For instance, deepfake technology allows bad actors to produce fake videos\nand audio clips of politicians and celebrities, manipulating their words and actions to deceive the public and sow\nconfusion. An example of this includes DRAGONBRIDGE's attempt to use AI-generated images to discredit U.S.\nleaders[^21]. However, such\ncampaigns have so far seen limited success.\n\nGenerative AI models can also be employed to craft realistic disinformation campaigns that undermine trust in\ninstitutions, destabilise social cohesion, and disrupt democratic processes. For example, during election periods, a\nsophisticated AI-generated disinformation campaign could manipulate public discourse by disseminating false news\nstories, conspiracy theories, and divisive content. The consequences of such actions can be far-reaching, swaying public\nopinion based on lies, and eroding trust in the democratic process.\n\nThe fact that these disinformation campaigns can be pre-planned and automated exacerbates the problem significantly,\nallowing malicious actors to generate and disseminate false information at an overwhelming scale. This makes it\nextremely challenging for fact-checkers, journalists, and social media platforms to identify and counteract the spread\nof disinformation in a timely manner. Moreover, the speed and efficiency with which AI can produce content makes it\nharder for users to distinguish between legitimate and fake news, further facilitating the spread of misinformation.\n\n### Censorship and control\n\nLarge AI models can also be exploited by authoritarian governments to manipulate public opinion and suppress democratic\nprocesses. By using these advanced technologies to generate fake news, propaganda, and deepfake content, such regimes\ncan craft an illusion of reality that aligns with their interests. This disinformation can cause confusion and distrust\namong the public, undermining the credibility of democratic institutions and opposition leaders.\n\nAdditionally, authoritarian governments can utilise AI-powered surveillance systems to track and monitor the activities\nof opposition members and dissidents. By analysing vast amounts of data from social media, communications, and location\ntracking, these models can create detailed profiles of individuals considered threats to the regime. The authorities can\nthen use this information to suppress dissenting voices through harassment, arrests, or other forms of repression.\n\n\n\n\n\n\n\n\n\n\nGenerative AI technology has emerged as a transformative innovation with the potential to disrupt industries and reshape\nsociety. While predicting its future trajectory is challenging due to the rapid pace of technological evolution, past\ntrends provide valuable insights. The launch of ChatGPT marked a pivotal moment, sparking widespread interest and\nchanging the way businesses and individuals interact with AI. This prompted competitors like Google and Anthropic to\nrelease proprietary models, although these remain tightly controlled by their developers. At the same time, the rise of\nopen-source models, such as those based on Meta's LLaMA or Mistral AI, has democratised access, allowing organisations\nand\nindividuals to deploy, customise, and run AI tools independently at lower costs. These models now rival their\nclosed-source counterparts in performance, providing privacy-conscious entities with a viable alternative for\non-premises deployment.\n\nHowever, this immense potential comes with inherent risks. The ability of Generative AI to create realistic content\nraises ethical concerns, including the proliferation of deepfakes, misinformation, and automated job displacement.\nBiases embedded in training data threaten to perpetuate discrimination, while the technology's rapid advancement poses\nchallenges for regulatory frameworks. Yet, Generative AI is not magic -- it builds on decades of research in machine\nlearning, neural networks, and computational power. Sophisticated algorithms process vast datasets to produce complex\noutputs, with today's state-of-the-art tools merely serving as a precursor to even more capable systems.\n\nAs progress in AI remains inevitable, organisations must proactively integrate these tools into their strategies while\nestablishing ethical guidelines and security protocols. The EU's \u20ac200 billion investment in AI infrastructure highlights\nthe urgency of this task, striving to balance innovation with accountability. Failing to engage with this shift risks\nceding competitive ground to those willing or maliciously inclined to exploit its benefits. Generative AI is here to\nstay, demanding a dual focus on harnessing its potential while mitigating its dangers.\n\n## Recommendations\n\nWe present recommendations to support Union entities in directing and coordinating their efforts concerning generative\nAI. Due to the rapidly evolving nature of this field, the recommendations are categorised into short-, medium-, and\nlong-term measures.\n\n### Short-term\n\n- **Stay informed on Generative AI developments**: Continuously monitor advancements in Generative AI, as these\n           technologies will likely impact various aspects of your operations and workflows.\n- **Invest in user awareness and training**: Promote responsible and informed usage of AI within your organisation by\n             ensuring that staff understand both the benefits and risks associated with the technology.\n- **Establish data handling policies**: Implement clear guidelines to ensure that only publicly available data (TLP:\n                CLEAR) is used in prompts submitted to commercial large language models, especially those provided online by public AI\n                services.\n- **Explore services under Cloud Broker PaaS Contracts**: Investigate privacy-enhanced commercial AI models, such as\n              Azure OpenAI or Mistral, which provide secure environments with regional hosting, ensuring user data stays within the\n              EU and isn't used for further training.\n- **Monitor open-source models**: Local, open-source models are progressing rapidly, with increasing potential for\n              customisation and fine-tuning on sensitive data. Keep track of these developments for future deployment opportunities.\n- **Engage with other institutions, such as CERT-EU, for expert guidance on securing your AI systems and mitigating\n    risks related to cybersecurity and data privacy.** CERT-EU's support can help ensure that your organisation follows\n    best practices in AI security.\n\n### Medium-term\n\n- **Develop a responsible AI policy**: Establish clear internal policies to guide the ethical and responsible use of\n              Generative AI technologies. Define acceptable use cases and implement validation processes for AI-generated outputs.\n- **Plan for local infrastructure deployment**: Consider deploying local, open-source Generative AI models either\n           on-premises or within private cloud environments. This would provide enhanced control over sensitive data while\n           benefiting from the growing capabilities of self-hosted models.\n- **Leverage the European Union's strategic AI initiatives**: With the EU committing to investments in AI and its\n               efforts to create an ecosystem that prioritises data sovereignty and security, consider aligning your AI strategy with\n               regional initiatives and future EU-backed cloud services to ensure compliance with privacy regulations.\n\n### Long-term\n\n- **Invest in advanced AI infrastructure**: As Generative AI continues to evolve, it will be crucial to invest in\n             scalable and resilient infrastructure to support complex AI workloads. Consider establishing partnerships with cloud\n             providers or with other institutions to ensure that your organisation can handle increasing data and processing\n             demands in a cost-effective and efficient manner.\n- **Foster AI-driven innovation**: Encourage research and innovation within your organisation by establishing AI-driven\n             labs or collaborating with other institutions such as CERT-EU. This can help create specialised models tailored to\n             your specific needs, advancing your competitiveness and strategic positioning in the AI space.\n- **Prepare for next-gen AI applications**: Look ahead to the next-generation AI technologies, such as AI systems with\n              advanced reasoning, multi-modal capabilities, and autonomous decision-making. Prepare your organisation to leverage\n              these innovations through long-term AI strategy planning, including upskilling your workforce and investing in future\n              technologies.\n\n\n\n\n\n\n\nWe would like to warmly thank our colleagues from European Commission for the useful input, feedback and suggestions they have provided to improve this guidance.\n\n\n\n\n\n\n\nIf you have suggestions that could help improve this document, please contact us at [services@cert.europa.eu](mailto:services@cert.europa.eu). We always appreciate constructive feedback.\n\n[^1]: _The Defender's Dilemma_: \"Defenders have to be right every time. Attackers only need to be right once.\"\n[^2]: [ENISA Threat Landscape 2024](https://www.enisa.europa.eu/sites/default/files/2024-11/ENISA%20Threat%20Landscape%202024_0.pdf)\n[^3]: [CSA State of the AI](https://services.google.com/fh/files/misc/csa_state_of_ai_and_security_survey_google_cloud.pdf)\n[^4]: [GPT@EC](https://commission.europa.eu/news/commission-launches-new-general-purpose-ai-tool-gptec-2024-10-22_en)\n[^5]: [EU launches InvestAI](https://ec.europa.eu/commission/presscorner/detail/en/ip_25_467)\n[^6]: [Microsoft Security Copilot](https://blogs.microsoft.com/blog/2023/03/28/introducing-microsoft-security-copilot-empowering-defenders-at-the-speed-of-ai/)\n[^7]: [New wave of AI powered capabilities](https://www.paloaltonetworks.com/blog/2023/06/a-new-wave-of-ai-powered-capabilities/)\n[^8]: [VirusTotal Code Insight](https://blog.virustotal.com/2023/04/introducing-virustotal-code-insight.html)\n[^9]: [GitHub Copilot](https://github.com/features/preview/copilot-x)\n[^10]: [Amazon CodeWhisperer](https://aws.amazon.com/codewhisperer/)\n[^11]: [Amazon Code Whisperer](https://aws.amazon.com/codewhisperer/)\n[^12]: [Microsoft 365 Copilot](https://blogs.microsoft.com/blog/2023/03/16/introducing-microsoft-365-copilot-your-copilot-for-work/)\n[^13]: [Google AI-powered workspace](https://workspace.google.com/blog/product-announcements/generative-ai)\n[^14]: [OpenAI terms of use](https://openai.com/policies/terms-of-use)\n[^15]: [OpenAI retention policy](https://help.openai.com/en/articles/7232945-how-can-i-use-the-chat-completion-api)\n[^16]: [Regulatory framework AI](https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai)\n[^17]: [On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?](https://dl.acm.org/doi/abs/10.1145/3442188.3445922)\n[^18]: [Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data](https://aclanthology.org/2020.acl-main.463.pdf)\n[^19]: [AI Act](https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai)\n[^20]: [How DeepSeek Censorship Actually Works](https://www.wired.com/story/deepseek-censorship/)\n[^21]: [Google disrupted over 10,000 instances of DRAGONBRIDGE activity](https://blog.google/threat-analysis-group/google-disrupted-dragonbridge-activity-q1-2024/)\n",
    "content_html": "<p><em><strong>Some portions of this document were crafted with a little help from our internally deployed AI models (powered by Llama 3.3 - 70B). While we have fine-tuned and polished the content, this also showcases how Generative AI can be a powerful tool -- especially when running on your own infrastructure for greater control, security and privacy.</strong></em></p><h3 id=\"generative-ai\">Generative AI</h3><p>Generative AI refers to a class of artificial intelligence models designed to simulate human-like creativity and adaptability by generating new content, data, or outputs based on learned patterns from vast datasets. These AI systems can be applied across a wide range of domains, from natural language processing to computer vision. For example, Generative AI can create realistic images, draft human-like text, compose music, or design novel chemical compounds. Notable examples include large language models (LLMs) such as OpenAI's GPT series, Mistral Le Chat, Google's Gemini, and Meta's LLaMA, as well as text-to-image generation tools like OpenAI's DALL-E and Stability AI's Stable Diffusion.</p><p>Large language models (LLMs) are advanced deep learning models that use self-attention mechanisms and multi-layered architectures to understand and generate text. These models excel in tasks such as language translation, summarisation, question-answering, and creative content generation by analysing vast quantities of data and identifying complex patterns. Their strengths lie in their ability to capture nuanced contextual information, generate coherent and relevant responses, and adapt to diverse domains.</p><p>However, LLMs also have notable limitations. They are highly data- and computationally intensive, requiring substantial resources for training and fine-tuning. Additionally, they may produce plausible-sounding yet incorrect or nonsensical answers (often referred to as hallucinations) and can be sensitive to input phrasing, leading to inconsistent results. Finally, they may inadvertently generate biased or harmful content due to biases present in their training data.</p><p>Similarly, text-to-image generation models utilise deep learning techniques to create visually coherent images based on textual input. The strengths of these models include their ability to generate diverse and creative images, as well as contributing to data augmentation and visual storytelling. However, like LLMs, their weaknesses include a dependence on large, well-annotated datasets for training, high computational requirements, and the potential to generate unrealistic or low-quality images. Furthermore, these models may struggle to accurately capture complex and abstract concepts described in the textual input and, similar to transformers, may inadvertently propagate biases present in the training data.</p><h3 id=\"how-does-it-work\">How does it work?</h3><p>Generative AI might seem like magic at first, but it's actually the result of significant progress in deep learning. This progress is driven by rapid increases in computing power, access to large datasets, and better training techniques such as reinforcement learning and self-supervised learning. These advances have come together to make Generative AI a reality, rather than just an idea.</p><h4 id=\"text-generation-models\">Text generation models</h4><p>At the core of large language models are neural networks with millions or even billions of parameters (i.e., <em>model size</em>), which are trained on vast amounts of text data. These <em>parameters</em> define the connections between the nodes in the network. Before an LLM can be used, it must undergo a training process, during which it is presented with massive datasets (known as training sets) that allow the model to learn patterns, relationships, and structures within the language. Through this, the model adjusts its parameters to minimise errors in prediction. Modern LLMs are trained using unsupervised learning and self-supervised learning, where the model learns by predicting the next word (or more precisely: <em>token</em>) in a sequence based on the prior context. LLMs can be broadly categorised into:</p><ul><li><p><strong>Foundational models</strong></p><p>Foundational models, such as OpenAI's GPT series, Anthropic's Claude, Meta's LLaMA, and others -- serve as the base for a wide range of applications. These models are pre-trained on enormous, diverse datasets sourced from books, articles, websites, and other publicly available texts. Rather than learning specific tasks, foundational models learn a probabilistic distribution of language -- in other words, they grasp how words, phrases, and concepts typically relate to one another across a wide array of contexts. They are general-purpose models capable of understanding and generating text across numerous domains without being tailored to specific tasks. Foundational models are highly versatile and can be applied to various tasks, including translation, summarisation, creative writing, and even code generation.</p></li><li><p><strong>Fine-tuned models</strong></p><p>Fine-tuned models build upon foundational models but are adapted for specific tasks or domains. They undergo additional training, or fine-tuning, using task-specific data or techniques such as Reinforcement Learning from Human Feedback (RLHF). Fine-tuning enhances the model's performance in particular applications, allowing it to better meet user needs. For instance, OpenAI's ChatGPT (based on the GPT-4 model) is a fine-tuned version of a foundational model, optimised for conversational AI. It has been trained not only to generate text but also to handle dialogue, ensuring that responses are contextually relevant, informative, and aligned with user intent. This fine-tuning process helps reduce errors and improves the model's reliability in real-world applications, such as customer service, virtual assistants, and personalised content generation.</p></li></ul><p>Once training is complete, the model enters the <em>inference</em> phase, where it generates predictions or completes tasks based on new input data. During this stage, the model leverages its internal knowledge of language patterns and relationships, acquired during training, to produce relevant and coherent output. For foundational models, this means generating responses based on a broad understanding of language and context, while fine-tuned models use additional task-specific training to generate more tailored and accurate responses.</p><h4 id=\"visual-generation-models\">Visual generation models</h4><p>Generative AI models are used to create visual content, such as images and videos, from textual descriptions. These models use neural networks to produce high-quality visual outputs. They are changing the creative industries by allowing dynamic content to be generated, whether it's a single image or a video sequence.</p><p>There are different types of models used for image and video generation. For example, Generative Adversarial Networks (GANs) and Diffusion Models, like DALL-E 3 and Stable Diffusion, are commonly used. Video models are more complex because they need to maintain both spatial and temporal coherence.</p><p>Image generation models focus on converting text prompts into static images. Diffusion Models have become popular in recent years and work by refining random noise into structured images using text prompts. To train these models, large datasets of paired text and image data are needed. This allows the model to learn relationships between textual descriptions and their corresponding visual representations.</p><p>Video generation builds on the foundations of image generation but introduces the challenge of spatio-temporal relationships. This means ensuring that individual frames and transitions between them are coherent over time. Video generation models often adapt Diffusion Models to handle sequential frame generation while considering temporal flow. They need to learn to generate high-quality images for each frame and ensure movements, lighting, and objects remain consistent throughout the sequence.</p><p>Both image and video generation models rely on the concept of latent space during inference. Latent space is an abstract representation of the learned relationships between visual and textual elements. In image generation, the model samples from this latent space to produce a single visual output. For video generation, the latent space also encodes temporal dynamics, enabling the model to generate a sequence of frames that align with the input text while ensuring smooth transitions.</p><p>To achieve this, video generation models are trained on vast video-text paired datasets. The focus is on learning to generate not just realistic images but also seamless motion and narrative progression. This results in dynamic content that feels natural while being driven by the input text. By understanding how these models work, you can explore their potential applications in various industries and creative projects.</p><h3 id=\"future-outlook\">Future outlook</h3><p>Artificial intelligence is advancing rapidly, with transformative innovations reshaping industries and unlocking new opportunities. The growth of open-source and open-weights models has significantly expanded AI's accessibility and applications. Unlike proprietary systems, these frameworks allow organisations, researchers, and developers to deploy, adapt, and refine AI tools without restrictive licensing, promoting greater autonomy. This shift is particularly evident in Generative AI, where advanced open source projects such as DeepSeek are driving significant breakthroughs. These models now enable diverse applications, from language processing to image and video generation, empowering businesses and individuals to innovate cost-effectively.</p><p>Cost-effective AI models are now rivaling proprietary models in performance. Research into self-supervised, unsupervised, and reinforcement learning is advancing rapidly, while breakthroughs in multimodal AI -- integrating text, images, audio, video, and even interactive environments -- are pushing the boundaries of creativity and problem-solving.</p><p>At CERT-EU, we are pursuing in-house AI projects to optimise operations and deepen our expertise in these technologies. This hands-on approach ensures that our advice on AI systems is both informed and actionable. While AI's democratisation offers opportunities to enhance innovation, personalise experiences, and automate tasks, it also presents significant risks, such as the spread of disinformation, misuse by malicious actors, and ethical dilemmas surrounding synthetic content. As the technology evolves, it is crucial to address these challenges proactively to ensure its responsible development.</p><p>Generative AI holds immense transformative potential, reshaping various sectors, including cybersecurity. This document specifically focuses on its implications for Union entities, examining how Generative AI is revolutionising both the defensive and offensive aspects of cybersecurity. By exploring how these technologies can strengthen protective measures while also enabling new threats, we aim to assess their impact on organisations within the Union.</p><p>Our goal is to propose actionable recommendations that will help direct and coordinate the efforts of Union entities in effectively harnessing the benefits of Generative AI, while mitigating its associated risks. Given the rapid pace of innovation, these insights reflect the landscape as of mid-2025, with the understanding that ongoing developments may substantially alter the threat environment.</p><h3 id=\"defensive-applications-of-ai\">Defensive applications of AI</h3><p>Artificial intelligence offers powerful tools to counter sophisticated cyber threats, mitigating the traditional \u201cDefender's Dilemma\u201d<sup class=\"footnote-ref\" id=\"fnref-1\"><a href=\"#fn-1\">1</a></sup> -- where attackers historically retain the upper hand. AI systems excel at transforming vast datasets into actionable intelligence, bolstering capabilities such as malware detection, vulnerability identification, and threat analysis. By automating routine tasks and accelerating response times, these technologies enable security teams to operate with heightened efficiency and precision.</p><p>Generative AI further augments defences. For example, it can simulate realistic cyber-attack scenarios, such as phishing campaigns or ransomware simulations, to rigorously test and train personnel. Beyond training, generative models facilitate the creation of adaptive honeypots that mislead attackers while gathering tactical intelligence. Large language models add value by detecting subtle patterns in data, such as log file anomalies, empowering analysts to prioritise risks and uncover hidden correlations.</p><h3 id=\"offensive-applications-of-ai\">Offensive applications of AI</h3><p>The same technologies empowering defenders are increasingly weaponised by adversaries. Generative AI, for instance, is exploited to craft hyper-realistic social engineering campaigns, including personalised phishing emails, SMS scams, and deepfake audio/video. Beyond deception, these models automate the discovery of software vulnerabilities -- even uncovering novel attack vectors, and generate functional code for malware or evasion techniques<sup class=\"footnote-ref\" id=\"fnref-2\"><a href=\"#fn-2\">2</a></sup>. Tools like WormGPT or FraudGPT, for example, illustrate how readily AI can be adapted to scale malicious activities.</p><p>The escalation of AI-driven disinformation and cyberattacks presents acute challenges for the Union entities. This evolving threat landscape underscores the urgent need for proactive, AI-enhanced security strategies, such as behaviour-based anomaly detection and predictive threat modelling, to counter adversarial innovation.</p><p>It is essential for CERT-EU to communicate not only the risks posed by Generative AI but also the opportunities it offers to enhance the resilience and cybersecurity capabilities of Union entities. Collaboration among these entities will be crucial for sharing best practices, investing in development, and establishing ethical frameworks for AI use. By prioritising human-AI collaboration and maintaining rigorous oversight, CERT-EU and Union entities can set a benchmark, fostering a secure and innovative digital ecosystem within the public institutions of the European Union.</p><p>As AI continues to evolve, organisations are now focused on integrating these technologies into their existing cybersecurity teams and systems. The challenge lies in ensuring the effective adoption of AI solutions, aligning them with strategic objectives, and optimising overall security efforts.</p><p>A 2024 survey by the Cloud Security Alliance (CSA)<sup class=\"footnote-ref\" id=\"fnref-3\"><a href=\"#fn-3\">3</a></sup> found that 63% of security professionals believe AI will enhance security measures. However, adoption remains in its early stages, with only 22% of organisations currently using Generative AI, though 55% plan to implement it within the next year.</p><h3 id=\"ais-impact-on-security-operations\">AI's impact on security operations</h3><p>AI-driven automation is reshaping security operations by improving threat detection, accelerating incident response, and reducing operational burdens. While concerns about AI replacing human roles exist, most professionals see it as a complementary tool. Only 12% believe AI will fully replace their roles, whereas 30% expect it to enhance their skills, and 28% see AI as broadly supporting their work. However, 51% warn against over-reliance, emphasising the need for human oversight in AI-driven security strategies.</p><p>Organisations are deploying AI to address workforce shortages and improve security efficiency. Around 36% of security teams use AI to bridge skill gaps, while 26% prioritise faster threat detection. Other key objectives include improving productivity and reducing misconfigurations. Generative AI is increasingly used for automated rule creation, attack simulations, and compliance monitoring, but its adoption introduces risks such as data manipulation and adversarial attacks. Effective governance frameworks are necessary to ensure responsible AI deployment.</p><h3 id=\"challenges-in-ai-adoption\">Challenges in AI adoption</h3><p>Despite its advantages, AI adoption in cybersecurity presents significant challenges. A primary concern is the shortage of skilled professionals who can effectively manage and secure AI systems, with 33% of organisations citing skill gaps as a major barrier. The complexity of AI models requires specialised expertise to train, maintain, and interpret their outputs. Additionally, 38% of security professionals highlight data quality issues, including unintended bias in AI models, while 25% cite privacy risks as a growing concern.</p><p>AI security risks extend beyond implementation challenges. Security professionals are increasingly aware of vulnerabilities such as data poisoning, adversarial attacks, and AI-generated misinformation. Approximately 28% of respondents express concerns about data poisoning attacks, which could manipulate AI models to generate false outputs. Regulatory and compliance considerations further complicate AI deployment, requiring organisations to navigate evolving security frameworks.</p><h3 id=\"future-trends-in-ai-driven-cybersecurity\">Future trends in AI-Driven cybersecurity</h3><p>AI adoption is expected to accelerate, with organisations increasingly exploring Generative AI for cybersecurity applications. Automated rule creation, attack simulations, and compliance monitoring are among the most common use cases. While AI-driven automation will continue to expand, security teams must balance efficiency gains with oversight mechanisms to mitigate potential risks.</p><p>Governance structures are evolving to support responsible AI deployment. Many organisations are establishing dedicated teams to oversee AI implementations, ensuring compliance with security policies and regulatory requirements. Executive leadership plays a crucial role in AI adoption, with 82% of organisations reporting strong leadership support for AI initiatives. However, a gap remains between executive-level strategy and operational execution, reinforcing the need for clear guidelines and AI-specific training programmes.</p><p>As AI adoption progresses, cybersecurity strategies must adapt to emerging threats. The increasing sophistication of AI-driven cyberattacks necessitates continuous advancements in AI-based defence mechanisms. Security teams must also address AI security risks, improve transparency in AI decision-making, and implement safeguards against adversarial manipulation. The widespread integration of AI into cybersecurity marks a transformative shift, requiring a strategic approach to maximise its benefits while mitigating associated risks.</p><h3 id=\"european-union-initiatives-and-investments-in-ai\">European Union initiatives and investments in AI</h3><p>The European Commission has made significant strides in promoting the secure and compliant use of Generative AI technologies. A notable example is its approval of the Azure OpenAI service for handling Sensitive Non-Classified Data, which ensures that user prompts remain within the EU and are excluded from model training. Hosted in regional cloud environments under Cloud Broker Contracts, this service strikes a balance by providing robust privacy safeguards without the complexities and resource-intensive demands of managing local models.</p><p>Similarly, the Commission has implemented GPT@EC<sup class=\"footnote-ref\" id=\"fnref-4\"><a href=\"#fn-4\">4</a></sup>, deploying AI models within a secure, local infrastructure to enhance operational efficiency while adhering to EU data protection standards. Complementing these efforts, CERT-EU has integrated internal large language models to power AI-driven projects, ensuring sensitive data remains protected while boosting productivity and supporting internal initiatives to reduce manual workloads and enhance decision-making processes</p><p>Looking ahead, the European Union has pledged \u20ac200 billion in AI investments<sup class=\"footnote-ref\" id=\"fnref-5\"><a href=\"#fn-5\">5</a></sup>, aiming to build critical infrastructure such as AI training gigafactories within the region. This ambitious plan highlights Europe's commitment to advancing AI innovation while maintaining a strong focus on data privacy and regulatory compliance.</p><p>As this document focuses on the impact of Generative AI on cybersecurity and its relevance to Union entities, several updated use cases in cybersecurity are highlighted, with an emphasis on ongoing AI-driven initiatives.</p><h3 id=\"improving-threat-detection\">Improving threat detection</h3><p>AI systems enhance cybersecurity by refining detection algorithms and generating new rules based on the latest threat intelligence. By leveraging AI-driven insights, security teams can improve detection capabilities, strengthening defences against cyberattacks.</p><p>Log analysis is a critical aspect of a cybersecurity analyst's role, and large language models (LLMs) can significantly streamline this process. AI can sift through vast amounts of log data, identifying anomalies, outliers, and correlations that may indicate security breaches. Automating log analysis reduces the time and effort required, allowing analysts to focus on high-priority threats rather than manually searching for patterns. Additionally, AI systems can detect connections between seemingly unrelated events, providing a comprehensive view of the security landscape and enabling more effective threat responses.</p><p>AI also holds promise in enhancing detection rules by analysing large datasets, such as system logs, to uncover patterns that may be missed by human analysts. However, false positives remain a major concern. Even with a low false positive rate of 0.1%, the sheer volume of log entries -- often in the hundreds of millions -- can result in an overwhelming number of alerts, necessitating careful tuning of AI models.</p><p>CERT-EU is exploring the use of retrieval-augmented generation (RAG)-based systems to assist analysts in creating and refining detection rules. These systems leverage existing rules to generate new ones, which analysts can review and implement. Other AI-driven platforms already being used for threat detection and response include Microsoft Security Copilot<sup class=\"footnote-ref\" id=\"fnref-6\"><a href=\"#fn-6\">6</a></sup> and Palo Alto Networks' Cortex XSOAR<sup class=\"footnote-ref\" id=\"fnref-7\"><a href=\"#fn-7\">7</a></sup>, demonstrating AI's growing role in strengthening cybersecurity operations.</p><h3 id=\"supporting-analysis\">Supporting analysis</h3><p>LLMs have also enabled analysts to work more efficiently and accurately in various aspects of their jobs. One key area where they have demonstrated their potential is in the deobfuscation of malicious code. Attackers often obfuscate their code to evade detection, but LLMs can assist analysts in identifying patterns and decoding hidden algorithms, providing valuable insights into the attacker's intent and revealing the true nature of the threat, thereby significantly speeding up the investigation.</p><p>Additionally, LLM transformers show a remarkable ability to correlate data from various sources and fields as they have been trained on diverse datasets. This vast training corpus enables the AI model to extract and synthesise information from various seemingly unrelated sources. By leveraging its deep learning capabilities, LLMs can then identify connections, patterns, and insights across these different sources. As a result, the AI model has proven to be a valuable tool in solving problems, providing novel insights, and identifying correlations that would otherwise be easy to miss.</p><p>Code analysis and reverse engineering are also areas where LLMs can provide substantial assistance. With their extensive knowledge base, LLMs can evaluate software code and explain its operation. In the case of reverse engineering, LLMs can help dissect complex obfuscated code and provide insights into its functionality and purpose. By understanding how a piece of malware or exploit operates, analysts can develop effective countermeasures to protect their systems and networks, while also saving time during the investigation.</p><p>For example, VirusTotal's<sup class=\"footnote-ref\" id=\"fnref-8\"><a href=\"#fn-8\">8</a></sup> integration of the Sec-PaLM model now generates natural language summaries of code snippets, enabling analysts to identify potential vulnerabilities more quickly. Similar features are likely to be integrated into many reverse engineering, sandboxes, and analysis tools in the future.</p><h3 id=\"automating-threat-intelligence\">Automating threat intelligence</h3><p>LLM transformers can greatly enhance the process of generating threat intelligence reports by automating the collection, analysis, and summarisation of relevant data. This not only saves time and effort, but it also ensures that the information presented to cybersecurity teams is accurate, up-to-date, and easily digestible. Armed with this intelligence, defenders can make more informed decisions and take proactive measures to protect their systems.</p><h3 id=\"coding-and-documentation\">Coding and documentation</h3><p>LLM transformers have already shown their impact on the field of software development. These AI models assist developers with code completion, bug detection, security vulnerability identification, and automatic documentation. By suggesting secure code snippets and identifying flaws early, LLMs help accelerate the development process and improve security.</p><p>Tools like GitHub Copilot X<sup class=\"footnote-ref\" id=\"fnref-9\"><a href=\"#fn-9\">9</a></sup> or Amazon CodeWhisperer<sup class=\"footnote-ref\" id=\"fnref-11\"><a href=\"#fn-11\">10</a></sup> are enhancing developer productivity and integrating security best practices into workflows. While not replacing human developers, these tools provide significant support, especially in preventing common vulnerabilities such as SQL injection or cross-site scripting.</p><p>Generative AI also automates routine tasks like writing documentation and generating unit tests, enabling developers to focus on more complex security issues. However, since LLM-generated code is based on public datasets, there is a risk of inheriting bugs or vulnerabilities, making human oversight essential to ensure secure, production-ready code.</p><p>For example, CERT-EU has deployed internal copilot tools that leverage the use of internal AI models to assist developers in writing secure code and identifying potential vulnerabilities. These tools have significantly improved the quality of code produced and reduced the time spent on manual code reviews.</p><h3 id=\"enhancing-cybersecurity-training\">Enhancing cybersecurity training</h3><p>The sophisticated natural language processing capabilities of large language models are playing an increasingly important role in upskilling cybersecurity personnel. AI-powered systems now offer contextualised explanations of complex cybersecurity concepts, enabling junior staff to bridge knowledge gaps more quickly. For example, AI platforms can provide personalised guidance on identifying and responding to emerging threats. By offering real-time support, these models help less experienced team members contribute more effectively to threat analysis and mitigation, thereby enhancing the overall efficiency of security teams.</p><p>Moreover, many institutions are now offering more frequent internal training sessions to help staff better understand and properly utilise AI tools. These training programmes are designed to ensure that teams can leverage AI tools effectively, enhancing their ability to respond to cybersecurity challenges in a rapidly evolving landscape.</p><h3 id=\"content-generation\">Content generation</h3><p>Finally, one of the most evident applications of Generative AI is the creation of high-quality content across a range of domains, including the automatic generation of technical documentation, corporate communications, and presentations. The ability of AI-driven content generation platforms to understand context and produce human-like text can enhance organisations\u2019 approaches to content creation. The anticipated widespread adoption of Generative AI is largely due to its impressive capacity to save time, reduce costs, and increase overall efficiency in producing a variety of content types.</p><p>In the field of cybersecurity, Generative AI can be employed to draft post-incident analyses, summarise threat intelligence feeds, and produce tailored security advisories. It also supports the automation of incident reports, enabling security teams to focus on more strategic tasks. By streamlining the drafting process, organisations can ensure that critical information is communicated clearly and promptly, facilitating quicker and more effective responses to emerging threats.</p><p>Similarly, AI-driven content generation can enhance the quality and efficiency of presentations. By leveraging data, Generative AI can dynamically generate visuals, suggest relevant talking points, and recommend persuasive storytelling techniques to engage audiences. This not only simplifies the process of creating presentations but also increases their overall impact and effectiveness.</p><p>Similarly to code generation examples, there are already several products either available or being rolled out that propose such AI-based enhancements and solutions. These include, for instance, Microsoft 365 Copilot<sup class=\"footnote-ref\" id=\"fnref-12\"><a href=\"#fn-12\">11</a></sup> and Google AI-powered Workspace Features<sup class=\"footnote-ref\" id=\"fnref-13\"><a href=\"#fn-13\">12</a></sup>.</p><p>Generative AI models usually require significant computational resources to function effectively. Choosing the right infrastructure is crucial to balancing performance, scalability, cost, and security. There are several deployment options available, each with distinct advantages and trade-offs.</p><h3 id=\"public-closed-source-models-paid-or-free\">Public closed-source models \u2013 paid or free</h3><p>The AI industry has seen a surge in public closed-source models offered by major tech companies. These models provide powerful AI capabilities but often function as black-box solutions, limiting user control over data processing and storage.</p><p>The dominant market model today revolves around \"free\" closed-source AI services, such as ChatGPT, DALL-E, Midjourney, and Google Gemini. These platforms make AI highly accessible but come with significant data privacy concerns. Their terms of use often indicate that input and output data may be stored outside the EU and could be used for further training and fine-tuning of the models. As a result, organisations handling sensitive information must assume that any data provided through these services could become public knowledge.</p><p>Additionally, the emergence of cost-effective AI models from competitors like China's DeepSeek has intensified market competition, prompting tech giants to reassess their offerings and pricing structures. While these services offer ease of use and accessibility, organisations must carefully evaluate whether the trade-offs in data sovereignty and compliance align with their operational and regulatory needs.</p><h3 id=\"locally-hosted-open-source-models\">Locally-hosted open-source models</h3><p>Deploying open-source AI models on local infrastructure has become a key strategy for organisations prioritising data control, security, and customisation. Models like LLaMA 3 and Mistral can be hosted on-premises or in private clouds, ensuring compliance with data sovereignty regulations while avoiding reliance on external providers.</p><p>However, maintaining these systems requires high-performance GPUs and skilled personnel, leading to the growing adoption of AI colocation services. These services provide access to cutting-edge computing infrastructure without the overhead of in-house data centre management.</p><p>As AI advances, organisations are increasingly balancing performance, cost, and security by adopting self-hosted open-source models and colocation solutions, ensuring greater control over AI applications while maintaining scalability.</p><h3 id=\"privacy-focused-commercial-closed-source-models-with-specific-conditions-of-use\">Privacy-focused commercial closed-source models with specific conditions of use</h3><p>In response to increasing concerns over data privacy and regulatory compliance, several tech companies have introduced privacy-focused AI services that adhere to strict data handling policies. These cloud providers now offer solutions that ensure user data remains within specified regions and is not used for training or fine-tuning models, striking a balance between leveraging advanced AI capabilities and maintaining control over sensitive data.</p><p>For organisations handling sensitive information, these services offer an opportunity to benefit from AI technologies while ensuring privacy compliance. Additionally, privacy-focused commercial models with negotiated terms of use are gaining traction. These models come with specific configurations and agreements that differ from public closed-source models, providing stronger data protection but imposing more stringent conditions for non-compliance. As these offerings become more prevalent, organisations must carefully assess the terms of service to ensure they align with privacy requirements and legal obligations within the evolving regulatory landscape.</p><p>When selecting a deployment option for Generative AI models, organisations must carefully evaluate their specific needs and privacy requirements. One promising option for Union entities is the adoption of privacy-focused commercial models with customised terms of use. These models offer stronger data protection than public closed-source solutions, but with stricter conditions that may carry serious consequences for non-compliance.</p><p>Risks associated with the use of Generative AI can broadly be divided into two main categories:</p><ul><li>Risks related to its use within an organisation.</li><li>Risks resulting from its use by others, including malicious actors.</li></ul><h3 id=\"risks-of-using-generative-ai\">Risks of using Generative AI</h3><p>Specific risks may arise from the potential use of Generative AI technology by the staff of Union entities. As with its benefits, the emphasis remains firmly on cybersecurity, with a particular focus on these organisations.</p><h4 id=\"indirect-prompt-injection-attacks\">Indirect prompt-injection attacks</h4><p>As Generative AI technology evolves, new possibilities and risks emerge. The recent development of various plugins and interfaces for external data sources that can be used in conjunction with some of the large language models -- or even independent AI Agents -- increase their capabilities but also introduce new risks.</p><p>One of these risks is the possibility of indirect prompt-injection attacks. When models are able to use external data -- websites, documents, emails, etc., such external data may potentialy be under the control of malicious actors. This can allow an attacker to attempt to influence the model's output by carefully crafting their input or \"prompt\", often embedding hidden instructions or biases. The AI model then inadvertently generates output that could potentially spread misinformation, reveal sensitive information, or produce other undesirable outcomes. Despite the input appearing harmless or neutral to a human observer, it can result in manipulated outputs, thus presenting significant security concerns in the application of AI technologies.</p><p>Indirect prompt-injection attacks are already occurring in the wild, with various tools and setups allowing LLMs to access external data sources being used as vectors for these attacks. Examples include hidden text in web pages, inside documents or emails that are then provided as input to LLMs by unsuspecting users.</p><p>A significant challenge is that the existing defences are not currently equipped to effectively counter these attacks. The subtlety of the manipulation makes detection extremely difficult, especially as the injected prompts often appear harmless or neutral to human observers or are not easily visible at all. While it is possible to configure the models to ignore certain types of these attacks or specific prompts, there is no obvious way to create a permanent fix. Users should be cautious when using AI tools on any input that may have been subject to malicious modification (e.g., web pages, external documents, incoming emails, etc.).</p><h4 id=\"disclosure-of-sensitive-data\">Disclosure of sensitive data</h4><p>The use of freely available, closed-source AI language models, such as ChatGPT, poses potential risks to sensitive data submitted in user prompts. As users interact with the model, they may inadvertently input confidential or personally identifiable information (PII) while seeking assistance or answers. Since this information is typically stored to enable the model to process and generate responses, there is a risk that sensitive data could be exposed, either through data breaches or during the training of future versions of the AI models. Without proper data anonymisation and privacy safeguards in place, such information could be misused by unauthorised parties, leading to identity theft, financial fraud, or reputational damage for both individuals and organisations involved.</p><p>For instance, OpenAI\u2019s current terms of use<sup class=\"footnote-ref\" id=\"fnref-14\"><a href=\"#fn-14\">13</a></sup> specify that while OpenAI does not use API content to improve their services, they may use non-API content (i.e., prompts and outputs from ChatGPT) for this purpose. Therefore, if confidential or sensitive data is entered as part of a ChatGPT prompt, it could eventually be exposed. OpenAI states that requests submitted via their API will be stored for 30 days<sup class=\"footnote-ref\" id=\"fnref-15\"><a href=\"#fn-15\">14</a></sup> and not used for training. However, there is no guaranteed proof of compliance or transparency regarding OpenAI\u2019s future plans.</p><p>In the event of a cyberattack on the infrastructure of an AI language model, there is a significant risk of data leakage. Such a breach could expose sensitive and private user information, including personal details, confidential conversations, and intellectual property. The consequences of such exposure could be wide-ranging, including compromised privacy, loss of user trust, and potential legal ramifications.</p><p>Organisations must also be vigilant in how AI tools are deployed within their environments. Employees should be trained not to enter sensitive data into public models, and technical controls -- such as data masking, secure API gateways, and audit logging -- should be implemented where possible.</p><p>The European Commission\u2019s AI Act<sup class=\"footnote-ref\" id=\"fnref-16\"><a href=\"#fn-16\">15</a></sup> seeks to enforce strict standards around AI deployment and data protection. By complying with its provisions, organisations can reduce the risk of unauthorised access, enhance system transparency, and demonstrate accountability in how sensitive data is handled. This regulatory framework fosters public trust in AI technologies and encourages investment in privacy-preserving solutions.</p><h4 id=\"copyright-violations\">Copyright violations</h4><p>Generative AI technologies, such as text and image generation models, have raised concerns about potential copyright violations as they become increasingly adept at creating content that closely resembles human-authored works. In the realm of text generation, AI-powered tools can produce articles, stories, or even poetry, often blurring the lines between human creativity and synthetic, machine-generated output. This raises questions about the originality of the content and whether the AI system has unintentionally reproduced or closely mimicked copyrighted materials.</p><p>For instance, if a text-generation AI model creates a story that closely resembles a popular novel, the copyright holder of the original novel may claim infringement, arguing that the AI-generated work could be perceived as a derivative of their copyrighted material.</p><p>Similarly, image-generation models have the capability to create visually appealing artwork, designs, and even photorealistic images. These AI-generated images could infringe upon copyrighted visual content if they closely resemble existing works, such as paintings, photographs, or graphic designs. For example, if an image-generation AI model were to create an artwork strikingly similar to a famous painting, it could lead to copyright disputes between the original artist and the creator of the AI-generated piece. Moreover, these concerns extend to the potential appropriation of elements from multiple copyrighted works to create a new image, which could lead to multiple copyright violation claims.</p><p>In both cases, the increasing sophistication of Generative AI technologies complicates the legal landscape surrounding copyright protection, as it becomes more challenging to determine the true authorship and originality of content.</p><p>Additionally, in some instances, the models powering Generative AI technologies are known to be trained on copyrighted content without the explicit approval of the authors. This raises additional concerns, as the organisations behind these models could be held liable for potential copyright infringement. By using copyrighted material to train their AI systems, organisations may inadvertently propagate the unauthorised reproduction or adaptation of protected works, opening themselves up to potential litigation. As a result, there is a growing need for more robust and transparent content acquisition policies to ensure that the data used to train AI models is either appropriately licensed or falls under the scope of fair use.</p><h4 id=\"false-or-inaccurate-information\">False or inaccurate information</h4><p>AI language models have become increasingly adept at generating high-quality text. However, these models have flaws, and the risk of providing false or inaccurate information remains significant<sup class=\"footnote-ref\" id=\"fnref-17\"><a href=\"#fn-17\">16</a></sup>. As AI language models are trained on vast amounts of data from the internet, they are susceptible to absorbing and perpetuating the biases, misconceptions, and inaccuracies present in that data. It is also important not to confuse Natural Language Processing (NLP), which these models excel at, with Natural Language Understanding (NLU), a significant challenge in AI research. A system trained solely on form has a priori no way to learn meaning<sup class=\"footnote-ref\" id=\"fnref-18\"><a href=\"#fn-18\">17</a></sup>. Consequently, users of these models must be aware of the potential pitfalls and exercise critical thinking when interpreting generated text.</p><p>One primary concern with AI language models is the possibility of bias. As these models learn from the data they are trained on, any biases present in the training data are likely to be absorbed and perpetuated by the model. This could manifest as gender, racial, or political biases, among others, and can lead to the generation of text that is offensive or perpetuates harmful stereotypes. In some cases, these biases may even cause the AI to provide misleading or outright false information, leading users astray and potentially reinforcing pre-existing biases.</p><p>Similarly, when generating computer code in various programming languages, Large Language Models (LLMs) often provide code containing errors or security vulnerabilities. This is primarily due to the training data these models are exposed to, which may include a diverse array of programming languages, styles, and quality levels. As a result, LLM-generated code may not always adhere to best practices or conform to the latest security standards. Additionally, these models lack the inherent ability to understand the context or specific requirements of a given task, which may lead to the production of code that is unsuitable, flawed, or even dangerous. Therefore, it is crucial for developers to carefully review and validate any code generated by LLMs before incorporating it into their projects to mitigate potential risks and ensure the safety and integrity of their software applications.</p><p>Another concern is the phenomenon of \"hallucinations\" where AI language models generate text that appears plausible but is entirely fabricated or lacks a factual basis. These hallucinations can occur for various reasons, such as the model trying to fill in gaps in its knowledge or attempting to provide a coherent response to an ambiguous or unfamiliar prompt. While these hallucinations can sometimes be relatively harmless, in other instances, they can lead to the dissemination of false information or contribute to the spread of misinformation.</p><h4 id=\"hype-abuse\">Hype abuse</h4><p>The rapid advancements in Generative AI technology and the surrounding hype have led to a surge in public interest and adoption. While these innovations undoubtedly offer numerous benefits and transformative potential, the excitement can also lead to possible pitfalls. With increased hype, bad actors may exploit the situation by creating fake applications or investment schemes, capitalising on the popularity of recognisable AI brand names to deceive users and achieve malicious objectives.</p><p>One such pitfall is the emergence of fake ChatGPT apps on Android and iOS platforms. These counterfeit apps, disguised as popular AI language models, may carry out invasive data harvesting activities. Unsuspecting users who download and interact with these malicious apps may inadvertently expose their personal information, including messages, contacts, and browsing history. The harvested data can then be used for various nefarious purposes, such as identity theft, targeted advertising, or even extortion. This underscores the importance of exercising caution when downloading mobile applications and ensuring they originate from trusted sources and developers.</p><p>Another potential pitfall linked to the hype around Generative AI is the creation of fictitious cryptocurrency tokens using recognisable AI brand names. Bad actors may design and market these tokens to lure in unsuspecting investors, who may believe they are investing in a promising AI venture. Once the scammers accumulate a substantial amount of funds, they may disappear, leaving the investors with worthless tokens and significant financial losses. This highlights the need for investors to conduct thorough research and due diligence before committing to any investment, particularly in emerging technologies like AI and cryptocurrencies.</p><h4 id=\"over-relying-on-technology\">Over-relying on technology</h4><p>Over-relying on Generative AI technology presents several potential dangers that could significantly impact organisations. One key concern is the possible erosion of competence among staff. As AI systems become increasingly capable of handling tasks traditionally carried out by humans, employees may become more reliant on these technologies. This dependence could lead to a decline in critical thinking and problem-solving skills, making staff less versatile and adaptive when faced with new challenges. Furthermore, as AI takes over routine tasks, workers may lose the ability to perform these manually, resulting in a loss of valuable expertise.</p><p>Another issue is the overconfidence in the quality of output generated by AI. Due to inherent limitations in AI models, such as the token limits that restrict the amount of information a language model can \"remember\", the generated content may not be as accurate, comprehensive, or contextually appropriate as users might expect. This could lead to situations where AI-generated content is accepted at face value, potentially causing misinformation or poorly informed decisions.</p><p>The over-reliance on AI technologies may also manifest as a failure to account for policy or political decisions that could limit their use. Governments and regulatory bodies are increasingly scrutinising the implications of AI on privacy, security, and social equality, as seen with regulations like the AI Act<sup class=\"footnote-ref\" id=\"fnref-19\"><a href=\"#fn-19\">18</a></sup>. Consequently, new policies or regulations may be introduced, imposing restrictions on the development, deployment, or use of AI technologies. Organisations that become overly dependent on AI systems may find themselves ill-prepared to adapt to these changes, leading to potential operational disruptions.</p><p>Finally, as highlighted in the benefits section, when using LLM tools for programming, it is essential to remember that the generated code may contain bugs or be insecure or unsuitable. Extra care must be taken when allowing staff and contractors to use LLMs for developing applications. The emphasis should be on ensuring thorough validation and testing of the generated code to mitigate potential risks.</p><h4 id=\"llms-opinions-advice-and-moral-values\">LLMs opinions, advice, and moral values</h4><p>Large Language Models (LLMs), such as ChatGPT or DeepSeek, should not be relied upon for opinions, advice, or moral guidance due to the inherent limitations in their design and the nature of their training data. While LLMs are powerful AI tools, they are not human beings with emotions, life experiences, or ethical systems. Instead, they are complex algorithms built to generate humanlike text based on patterns and associations identified within vast amounts of data.</p><p>One of the primary reasons LLMs are unsuitable for providing opinions, advice, or moral guidance is that their responses are derived from the datasets used in their training. These datasets consist of vast amounts of text from a wide variety of sources, which may contain conflicting opinions, values, and perspectives. When an LLM encounters such contradictions, it may struggle to produce a coherent and consistent response. As a result, the output may seem random, as the model attempts to balance opposing viewpoints or may simply reproduce popular opinions without understanding the underlying reasons or nuances.</p><p>Furthermore, LLMs are incapable of forming independent opinions or moral judgments. They do not have the capacity to critically analyse complex issues or empathise with human emotions, both of which are essential when providing sound advice or ethical guidance. Relying on an LLM for such matters could lead to misguided or superficial conclusions that fail to address the unique complexities of a given situation.</p><p>It is therefore not surprising that, for example, Deepseek has imposed strict guardrails on the values that LLMs must reflect, aligning them with those of the Chinese Communist Party, as part of its broader efforts to enforce censorship and control over AI-generated content<sup class=\"footnote-ref\" id=\"fnref-20\"><a href=\"#fn-20\">19</a></sup>. After all, an LLM will inevitably reflect the moral values embedded in its training data and shaped by human feedback (Reinforcement Learning with Human Feedback -- RLHF). Consequently, for any generated text intended for political purposes, it may be wise to verify whether it aligns with the general vision, policy, and strategy of Union entities.</p><h3 id=\"risks-from-adversarial-use-of-generative-ai-technology\">Risks from adversarial use of Generative AI technology</h3><p>Specific risks arise from the use of Generative AI technology by malicious actors. As previously mentioned, the focus remains on cybersecurity, particularly with regard to Union entities.</p><h4 id=\"privacy-issues\">Privacy issues</h4><p>Personally identifiable information (PII) can inadvertently be included in the training datasets of generative AI models when data is collected from a broad range of sources, such as websites, forums, social media, and other digital platforms. This data may not always be properly anonymised or sanitised before being used to train the models. As a result, PII \u2014 including names, addresses, telephone numbers, email addresses, or other sensitive details \u2014 could become embedded within the model\u2019s training data, potentially allowing it to be traced back to specific individuals.</p><p>When these models are deployed, there is a risk that PII could be unintentionally disclosed through generated outputs. However, malicious actors may also deliberately exploit generative AI models to extract or reconstruct sensitive information, using techniques such as prompt injection or model inversion. These attacks are designed to probe the model for private or confidential data that may have been memorised during training.</p><p>This presents a twofold privacy risk: firstly, the unauthorised disclosure or targeted extraction of sensitive information could have serious consequences for the individuals concerned; secondly, the generated content may be inaccurate or misleading, resulting in misinformation or the misidentification of individuals. Both scenarios pose significant threats to privacy, trust, and the safe adoption of AI technologies \u2014 particularly in sensitive sectors such as government, healthcare, and finance.</p><h4 id=\"more-advanced-cyberattacks\">More advanced cyberattacks</h4><p>Generative AI technologies could also give rise to new methods for conducting cyberattacks. As AI systems become more sophisticated, they can be increasingly exploited by malicious actors to facilitate attacks and exploit vulnerabilities in various ways.</p><p>One such method involves using AI to generate phishing content. By harnessing natural language processing and generation capabilities, cybercriminals can craft highly convincing emails, text messages, and social media posts that appear to come from legitimate sources. These AI-generated messages can be specifically tailored to target individuals, increasing the likelihood of them falling victim to the scam. Furthermore, AI can automate the process of sending phishing messages, allowing attackers to target a larger number of potential victims.</p><p>Social engineering attacks can also be enhanced by AI-generated voice and video deepfakes. These realistic forgeries can be used to impersonate executives, celebrities, or other influential figures, manipulating victims into providing sensitive information or taking actions that benefit the attacker. Deepfake technology can also be employed to create more convincing phone scams or video calls, further increasing the likelihood of a successful attack.</p><p>Additionally, AI technologies can be used to improve malware, making it more difficult to detect and more effective in its operations. For example, AI algorithms can analyse existing malware and identify patterns likely to be flagged by antivirus software. Based on this analysis, AI can then generate new, stealthier malware variants that are harder to detect and better exploit system vulnerabilities.</p><p>AI can also facilitate cyberattacks through more efficient vulnerability detection and fuzzing. By using AI-powered tools, attackers can rapidly discover security weaknesses in software or network infrastructure, much faster than traditional methods. This allows them to identify and exploit vulnerabilities before they are patched, increasing the likelihood of a successful attack.</p><p>Furthermore, AI can be used to automate and optimise password cracking. By employing machine learning algorithms, attackers can recognise patterns in password creation and generate more effective password dictionaries, significantly speeding up the cracking process. This can drastically reduce the time it takes to gain unauthorised access to accounts, making it more difficult for security professionals to respond effectively.</p><p>Finally, the development of freely available Generative AI tools has inadvertently lowered the entry barriers for new malicious actors in the cybercrime ecosystem. With minimal technical expertise, individuals can exploit the capabilities of advanced AI models to conduct various illicit activities, such as generating phishing emails, creating realistic deepfakes, or producing fake news. This democratisation of access to powerful AI-driven tools amplifies the reach and impact of cybercrime, cyber espionage, and other malicious activities. It also poses significant challenges for cybersecurity professionals, law enforcement, and policymakers, as it allows a wider range of actors to participate in these activities.</p><h4 id=\"disinformation\">Disinformation</h4><p>The powerful capabilities of Generative AI models come with significant risks when misused for disinformation campaigns. These models can impersonate public figures and create highly convincing narratives, making them potent tools for spreading false and misleading information. For instance, deepfake technology allows bad actors to produce fake videos and audio clips of politicians and celebrities, manipulating their words and actions to deceive the public and sow confusion. An example of this includes DRAGONBRIDGE's attempt to use AI-generated images to discredit U.S. leaders<sup class=\"footnote-ref\" id=\"fnref-21\"><a href=\"#fn-21\">20</a></sup>. However, such campaigns have so far seen limited success.</p><p>Generative AI models can also be employed to craft realistic disinformation campaigns that undermine trust in institutions, destabilise social cohesion, and disrupt democratic processes. For example, during election periods, a sophisticated AI-generated disinformation campaign could manipulate public discourse by disseminating false news stories, conspiracy theories, and divisive content. The consequences of such actions can be far-reaching, swaying public opinion based on lies, and eroding trust in the democratic process.</p><p>The fact that these disinformation campaigns can be pre-planned and automated exacerbates the problem significantly, allowing malicious actors to generate and disseminate false information at an overwhelming scale. This makes it extremely challenging for fact-checkers, journalists, and social media platforms to identify and counteract the spread of disinformation in a timely manner. Moreover, the speed and efficiency with which AI can produce content makes it harder for users to distinguish between legitimate and fake news, further facilitating the spread of misinformation.</p><h4 id=\"censorship-and-control\">Censorship and control</h4><p>Large AI models can also be exploited by authoritarian governments to manipulate public opinion and suppress democratic processes. By using these advanced technologies to generate fake news, propaganda, and deepfake content, such regimes can craft an illusion of reality that aligns with their interests. This disinformation can cause confusion and distrust among the public, undermining the credibility of democratic institutions and opposition leaders.</p><p>Additionally, authoritarian governments can utilise AI-powered surveillance systems to track and monitor the activities of opposition members and dissidents. By analysing vast amounts of data from social media, communications, and location tracking, these models can create detailed profiles of individuals considered threats to the regime. The authorities can then use this information to suppress dissenting voices through harassment, arrests, or other forms of repression.</p><p>Generative AI technology has emerged as a transformative innovation with the potential to disrupt industries and reshape society. While predicting its future trajectory is challenging due to the rapid pace of technological evolution, past trends provide valuable insights. The launch of ChatGPT marked a pivotal moment, sparking widespread interest and changing the way businesses and individuals interact with AI. This prompted competitors like Google and Anthropic to release proprietary models, although these remain tightly controlled by their developers. At the same time, the rise of open-source models, such as those based on Meta's LLaMA or Mistral AI, has democratised access, allowing organisations and individuals to deploy, customise, and run AI tools independently at lower costs. These models now rival their closed-source counterparts in performance, providing privacy-conscious entities with a viable alternative for on-premises deployment.</p><p>However, this immense potential comes with inherent risks. The ability of Generative AI to create realistic content raises ethical concerns, including the proliferation of deepfakes, misinformation, and automated job displacement. Biases embedded in training data threaten to perpetuate discrimination, while the technology's rapid advancement poses challenges for regulatory frameworks. Yet, Generative AI is not magic -- it builds on decades of research in machine learning, neural networks, and computational power. Sophisticated algorithms process vast datasets to produce complex outputs, with today's state-of-the-art tools merely serving as a precursor to even more capable systems.</p><p>As progress in AI remains inevitable, organisations must proactively integrate these tools into their strategies while establishing ethical guidelines and security protocols. The EU's \u20ac200 billion investment in AI infrastructure highlights the urgency of this task, striving to balance innovation with accountability. Failing to engage with this shift risks ceding competitive ground to those willing or maliciously inclined to exploit its benefits. Generative AI is here to stay, demanding a dual focus on harnessing its potential while mitigating its dangers.</p><h3 id=\"recommendations\">Recommendations</h3><p>We present recommendations to support Union entities in directing and coordinating their efforts concerning generative AI. Due to the rapidly evolving nature of this field, the recommendations are categorised into short-, medium-, and long-term measures.</p><h4 id=\"short-term\">Short-term</h4><ul><li><strong>Stay informed on Generative AI developments</strong>: Continuously monitor advancements in Generative AI, as these technologies will likely impact various aspects of your operations and workflows.</li><li><strong>Invest in user awareness and training</strong>: Promote responsible and informed usage of AI within your organisation by ensuring that staff understand both the benefits and risks associated with the technology.</li><li><strong>Establish data handling policies</strong>: Implement clear guidelines to ensure that only publicly available data (TLP: CLEAR) is used in prompts submitted to commercial large language models, especially those provided online by public AI services.</li><li><strong>Explore services under Cloud Broker PaaS Contracts</strong>: Investigate privacy-enhanced commercial AI models, such as Azure OpenAI or Mistral, which provide secure environments with regional hosting, ensuring user data stays within the EU and isn't used for further training.</li><li><strong>Monitor open-source models</strong>: Local, open-source models are progressing rapidly, with increasing potential for customisation and fine-tuning on sensitive data. Keep track of these developments for future deployment opportunities.</li><li><strong>Engage with other institutions, such as CERT-EU, for expert guidance on securing your AI systems and mitigating risks related to cybersecurity and data privacy.</strong> CERT-EU's support can help ensure that your organisation follows best practices in AI security.</li></ul><h4 id=\"medium-term\">Medium-term</h4><ul><li><strong>Develop a responsible AI policy</strong>: Establish clear internal policies to guide the ethical and responsible use of Generative AI technologies. Define acceptable use cases and implement validation processes for AI-generated outputs.</li><li><strong>Plan for local infrastructure deployment</strong>: Consider deploying local, open-source Generative AI models either on-premises or within private cloud environments. This would provide enhanced control over sensitive data while benefiting from the growing capabilities of self-hosted models.</li><li><strong>Leverage the European Union's strategic AI initiatives</strong>: With the EU committing to investments in AI and its efforts to create an ecosystem that prioritises data sovereignty and security, consider aligning your AI strategy with regional initiatives and future EU-backed cloud services to ensure compliance with privacy regulations.</li></ul><h4 id=\"long-term\">Long-term</h4><ul><li><strong>Invest in advanced AI infrastructure</strong>: As Generative AI continues to evolve, it will be crucial to invest in scalable and resilient infrastructure to support complex AI workloads. Consider establishing partnerships with cloud providers or with other institutions to ensure that your organisation can handle increasing data and processing demands in a cost-effective and efficient manner.</li><li><strong>Foster AI-driven innovation</strong>: Encourage research and innovation within your organisation by establishing AI-driven labs or collaborating with other institutions such as CERT-EU. This can help create specialised models tailored to your specific needs, advancing your competitiveness and strategic positioning in the AI space.</li><li><strong>Prepare for next-gen AI applications</strong>: Look ahead to the next-generation AI technologies, such as AI systems with advanced reasoning, multi-modal capabilities, and autonomous decision-making. Prepare your organisation to leverage these innovations through long-term AI strategy planning, including upskilling your workforce and investing in future technologies.</li></ul><p>We would like to warmly thank our colleagues from European Commission for the useful input, feedback and suggestions they have provided to improve this guidance.</p><p>If you have suggestions that could help improve this document, please contact us at <a rel=\"noopener\" target=\"_blank\" href=\"mailto:services@cert.europa.eu\">services@cert.europa.eu</a>. We always appreciate constructive feedback.</p><div class=\"footnotes\"><hr><ol><li id=\"fn-1\"><p><em>The Defender's Dilemma</em>: \"Defenders have to be right every time. Attackers only need to be right once.\"&#160;<a href=\"#fnref-1\" class=\"footnoteBackLink\" title=\"Jump back to footnote 1 in the text.\">&#8617;</a></p></li><li id=\"fn-2\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://www.enisa.europa.eu/sites/default/files/2024-11/ENISA%20Threat%20Landscape%202024_0.pdf\">ENISA Threat Landscape 2024</a>&#160;<a href=\"#fnref-2\" class=\"footnoteBackLink\" title=\"Jump back to footnote 2 in the text.\">&#8617;</a></p></li><li id=\"fn-3\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://services.google.com/fh/files/misc/csa_state_of_ai_and_security_survey_google_cloud.pdf\">CSA State of the AI</a>&#160;<a href=\"#fnref-3\" class=\"footnoteBackLink\" title=\"Jump back to footnote 3 in the text.\">&#8617;</a></p></li><li id=\"fn-4\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://commission.europa.eu/news/commission-launches-new-general-purpose-ai-tool-gptec-2024-10-22_en\">GPT@EC</a>&#160;<a href=\"#fnref-4\" class=\"footnoteBackLink\" title=\"Jump back to footnote 4 in the text.\">&#8617;</a></p></li><li id=\"fn-5\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://ec.europa.eu/commission/presscorner/detail/en/ip_25_467\">EU launches InvestAI</a>&#160;<a href=\"#fnref-5\" class=\"footnoteBackLink\" title=\"Jump back to footnote 5 in the text.\">&#8617;</a></p></li><li id=\"fn-6\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://blogs.microsoft.com/blog/2023/03/28/introducing-microsoft-security-copilot-empowering-defenders-at-the-speed-of-ai/\">Microsoft Security Copilot</a>&#160;<a href=\"#fnref-6\" class=\"footnoteBackLink\" title=\"Jump back to footnote 6 in the text.\">&#8617;</a></p></li><li id=\"fn-7\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://www.paloaltonetworks.com/blog/2023/06/a-new-wave-of-ai-powered-capabilities/\">New wave of AI powered capabilities</a>&#160;<a href=\"#fnref-7\" class=\"footnoteBackLink\" title=\"Jump back to footnote 7 in the text.\">&#8617;</a></p></li><li id=\"fn-8\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://blog.virustotal.com/2023/04/introducing-virustotal-code-insight.html\">VirusTotal Code Insight</a>&#160;<a href=\"#fnref-8\" class=\"footnoteBackLink\" title=\"Jump back to footnote 8 in the text.\">&#8617;</a></p></li><li id=\"fn-9\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://github.com/features/preview/copilot-x\">GitHub Copilot</a>&#160;<a href=\"#fnref-9\" class=\"footnoteBackLink\" title=\"Jump back to footnote 9 in the text.\">&#8617;</a></p></li><li id=\"fn-11\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://aws.amazon.com/codewhisperer/\">Amazon Code Whisperer</a>&#160;<a href=\"#fnref-11\" class=\"footnoteBackLink\" title=\"Jump back to footnote 10 in the text.\">&#8617;</a></p></li><li id=\"fn-12\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://blogs.microsoft.com/blog/2023/03/16/introducing-microsoft-365-copilot-your-copilot-for-work/\">Microsoft 365 Copilot</a>&#160;<a href=\"#fnref-12\" class=\"footnoteBackLink\" title=\"Jump back to footnote 11 in the text.\">&#8617;</a></p></li><li id=\"fn-13\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://workspace.google.com/blog/product-announcements/generative-ai\">Google AI-powered workspace</a>&#160;<a href=\"#fnref-13\" class=\"footnoteBackLink\" title=\"Jump back to footnote 12 in the text.\">&#8617;</a></p></li><li id=\"fn-14\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://openai.com/policies/terms-of-use\">OpenAI terms of use</a>&#160;<a href=\"#fnref-14\" class=\"footnoteBackLink\" title=\"Jump back to footnote 13 in the text.\">&#8617;</a></p></li><li id=\"fn-15\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://help.openai.com/en/articles/7232945-how-can-i-use-the-chat-completion-api\">OpenAI retention policy</a>&#160;<a href=\"#fnref-15\" class=\"footnoteBackLink\" title=\"Jump back to footnote 14 in the text.\">&#8617;</a></p></li><li id=\"fn-16\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai\">Regulatory framework AI</a>&#160;<a href=\"#fnref-16\" class=\"footnoteBackLink\" title=\"Jump back to footnote 15 in the text.\">&#8617;</a></p></li><li id=\"fn-17\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://dl.acm.org/doi/abs/10.1145/3442188.3445922\">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?</a>&#160;<a href=\"#fnref-17\" class=\"footnoteBackLink\" title=\"Jump back to footnote 16 in the text.\">&#8617;</a></p></li><li id=\"fn-18\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://aclanthology.org/2020.acl-main.463.pdf\">Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data</a>&#160;<a href=\"#fnref-18\" class=\"footnoteBackLink\" title=\"Jump back to footnote 17 in the text.\">&#8617;</a></p></li><li id=\"fn-19\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://digital-strategy.ec.europa.eu/en/policies/regulatory-framework-ai\">AI Act</a>&#160;<a href=\"#fnref-19\" class=\"footnoteBackLink\" title=\"Jump back to footnote 18 in the text.\">&#8617;</a></p></li><li id=\"fn-20\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://www.wired.com/story/deepseek-censorship/\">How DeepSeek Censorship Actually Works</a>&#160;<a href=\"#fnref-20\" class=\"footnoteBackLink\" title=\"Jump back to footnote 19 in the text.\">&#8617;</a></p></li><li id=\"fn-21\"><p><a rel=\"noopener\" target=\"_blank\" href=\"https://blog.google/threat-analysis-group/google-disrupted-dragonbridge-activity-q1-2024/\">Google disrupted over 10,000 instances of DRAGONBRIDGE activity</a>&#160;<a href=\"#fnref-21\" class=\"footnoteBackLink\" title=\"Jump back to footnote 20 in the text.\">&#8617;</a></p></li></ol></div>",
    "licence": {
        "title": "Creative Commons Attribution 4.0 International (CC-BY 4.0)",
        "link": "https://creativecommons.org/licenses/by/4.0/",
        "restrictions": "https://cert.europa.eu/legal-notice",
        "author": "The Cybersecurity Service for the Union institutions, bodies, offices and agencies"
    }
}