Advanced

GPAI / Foundation Model Rules

A practical guide to gpai / foundation model rules for AI engineers and policymakers.

What This Lesson Covers

GPAI / Foundation Model Rules is a key topic in EU AI Act Deep Dive. In this lesson you will learn the underlying concept, why it matters specifically for AI engineers and policymakers, the practical approach experienced teams use, and the patterns to avoid. By the end you will be able to engage with gpai / foundation model rules in real product and policy decisions.

This lesson belongs to the Regulation & Compliance category of the AI Ethics & Governance track. Ethics and governance are not optional add-ons — they shape what AI products are allowed to exist, what markets they can enter, and whether the underlying business model holds up under scrutiny.

Why It Matters

Master the EU AI Act. Learn the risk categories, prohibited practices, high-risk AI obligations, GPAI rules, timelines, penalties, and compliance roadmap.

The reason gpai / foundation model rules deserves dedicated attention is that the gap between AI teams that take ethics and governance seriously and those that don't is widening fast. Two teams shipping similar products can end up in very different positions when regulators, journalists, customers, or affected communities ask the hard questions. Ethics and governance done well are competitive advantages — not just compliance burdens.

💡
Mental model: Treat gpai / foundation model rules as a deliberate product and policy decision, not a checkbox. The teams shipping the most trustworthy AI weave ethics into engineering reviews, product roadmaps, and incident playbooks — not into a single offline document that no one reads.

How It Works in Practice

Below is a practical example of how to apply gpai / foundation model rules in real AI work. Read it once, then think about how you would adapt it to your specific product, regulatory environment, and stakeholders.

# EU AI Act risk categorization helper
EU_AI_ACT_CATEGORIES = {
    "unacceptable": [
        "social scoring by governments",
        "real-time remote biometric ID in public (with narrow exceptions)",
        "subliminal manipulation causing harm",
        "exploiting vulnerabilities of children/disabled",
        "emotion recognition in workplace and education",
    ],
    "high_risk": [
        "biometric categorization",
        "critical infrastructure (water, gas, electricity)",
        "education and vocational training (admissions, scoring)",
        "employment (hiring, performance, termination)",
        "essential services (credit scoring, benefits)",
        "law enforcement",
        "migration / border control",
        "administration of justice",
    ],
    "limited_risk":   ["chatbots, deepfakes (transparency obligations)"],
    "minimal_risk":   ["spam filters, AI in video games"],
    "gpai_systemic":  ["foundation models > 10^25 FLOPs (Article 51)"],
}

def classify_eu_ai_risk(use_case: str) -> str:
    for category, examples in EU_AI_ACT_CATEGORIES.items():
        if any(e in use_case.lower() for e in examples):
            return category
    return "minimal_risk"

# High-risk AI obligations: risk mgmt, data governance, technical docs,
# logging, transparency, human oversight, accuracy/robustness, post-market monitoring
# Penalties: up to EUR 35M or 7% of global annual turnover

Step-by-Step Walkthrough

  1. Identify the affected stakeholders — Not just users. Affected non-users, regulators, employees, and society at large all have stakes in AI decisions. Ethics is about who is in the room, not just whose voice is loudest.
  2. Ground the decision in a framework — Pick one: NIST AI RMF, ISO 42001, EU AI Act risk categorization, or your internal ethics framework. Ungrounded debate goes in circles.
  3. Get the inputs — Data on bias, customer feedback, regulator signals, comparable cases. Decisions made without inputs are guesses.
  4. Document the decision and the reasoning — Future-you and future regulators will want to know what you decided and why. Architecture Decision Records (ADRs) work well.
  5. Build in re-review cadence — Ethics norms shift faster than code. Set a calendar reminder to re-evaluate at 6 months, 12 months, and after every material change.

When To Use It (and When Not To)

GPAI / Foundation Model Rules applies when:

  • The AI feature touches people in consequential ways (jobs, money, freedom, health)
  • You operate in a regulated market or one likely to be regulated soon
  • The use case involves protected characteristics, vulnerable populations, or public interest
  • The cost of getting it wrong (in trust, lawsuits, or harm) outweighs the cost of doing it right

It is the wrong move when:

  • A simpler approach (a different feature, a different framing) avoids the ethics challenge entirely
  • You are still iterating on whether the feature should exist at all — decide that first
  • You are using ethics as a smokescreen to delay shipping a feature you privately know is fine
  • The decision is being made unilaterally by people without standing — pause and bring in the right voices
Common pitfall: Teams treat ethics review as a one-time approval rather than an ongoing operating practice. Norms shift, regulations change, and real-world impact often only becomes clear after deployment. Build the review cadence into your release process the way you build security review — not into a one-off document.

Practitioner Checklist

  • Have you identified all affected stakeholders, including non-users?
  • Is the decision grounded in a recognized framework (NIST, ISO, EU AI Act, internal)?
  • Have you measured the relevant fairness, privacy, transparency, and safety metrics?
  • Is there a documented decision record (ADR) with the reasoning, dissent, and alternatives?
  • Is there a plan to monitor real-world impact and re-evaluate?
  • Have you involved the right voices (legal, ethics, impacted communities, regulators where appropriate)?

Next Steps

The other lessons in EU AI Act Deep Dive build directly on this one. Once you are comfortable with gpai / foundation model rules, the natural next step is to combine it with the patterns in the surrounding lessons — that is where ethical practice goes from one-off decisions to an operating system. Ethics is most useful as a system, not as isolated reviews.