You are using an outdated browser. For a faster, safer browsing experience, upgrade for free today.

1. AI Foundations & Regulatory Context

Feedback | Download:
AI Foundations & Regulatory Context
1.1 An Introduction to AI for MSMEs
0%
Module Overview

This module provides an introductory overview of Artificial Intelligence (AI) tailored for micro, small, and medium-sized enterprises (MSMEs). The goal is to build a basic understanding of AI, its business relevance, key regulatory requirements, and ethical considerations. The content uses simple language, real-world MSME examples (e.g., retail shops, service providers, or small manufacturers), and avoids technical jargon. Explanations include analogies, such as comparing AI to a smart assistant, to make concepts accessible to all educational and professional backgrounds.

 

Knowledge, Skills, and Attitudes for Module 1
  • Understand basic AI definitions and concepts, including how AI differs from traditional tools; 
  • Articulate AI's potential business value through simple examples; 
  • Develop curiosity towards AI's capabilities while maintaining realistic expectations; 
  • Comprehend basic AI regulatory frameworks and their relevance to business operations; 
  • Identify common compliance requirements and apply basic checks (e.g., risk assessments); 
  • Prioritize ethical and legal considerations in AI adoption to build trust and avoid penalties;
  • Understand key principles of fairness, transparency, and privacy in AI use.
  • Identify responsible practices and apply simple safeguards (e.g., reviewing outputs, protecting sensitive data).
  • Value accountability and maintain trust in how AI is applied.

 

1.2 AI Fundamentals for Business
0%
Unit Overview:
In today's digital economy, AI helps MSMEs compete by automating tasks and providing insights.
   
AI adoption among European SMEs grew 20% in 2023, boosting efficiency (Organisation for Economic Co-operation and Development "SME Digitalisation" 2022).
   
This unit offers a foundational overview of AI, differentiating common myths from realities and highlighting its core value for businesses.
   
It uses everyday examples to demystify AI and show how it can support MSME operations like customer service or inventory management.
   
For instance, understanding AI basics will help a small retail MSME apply chatbots ethically, avoiding compliance pitfalls discussed later.

 

Understanding What is AI?
 
 

Definition:

AI as systems that mimic human intelligence to perform tasks like recognizing patterns or making predictions.

Analogy:

Compare AI to a recipe-following chef that learns from past meals.

Myths vs. Realities:

Myth—"AI will replace all jobs“.
Reality—AI augments human work, freeing time for creativity in MSMEs (Deloitte "Tech Trends" 2023).

AI as a solution:
Artificial intelligence can automate routine enquiries, reduce waiting times and thus relieve the burden on small teams.

 

 

 

What Is Artificial Intelligence?
ARTIFICIAL   INTELLIGENCE   ARTIFICAL INTELLIGENCE
Artificial intelligence is intelligence exhibited by machines, rather than humans or other animals. The field of AI research defines itself as the study of "intelligent agents": any device that perceives its environment and takes actions that maximize its chance of success at some goal.

 

History of Artificial Intelligence

History of Artificial Intelligence

1943: McCullough and Pitts' invented "artificial neurons".
1950: Inventions of Alan Turing's "Computing machinery and Intelligence".
1951: Al was using in Games.
1956: Dartmouth conference; And the birth of Al.
1965: Robinson's complete algorithm for logical reasoning.
1969-79: Early development of knowledge based systems took place.
1980-88: Expert system industry booms.
1988-93: Expert system industry busts "Al winter".
1993 - Present: Al is now using rapidly in different technologies; And is achieving it's goal.
2022: Introduction of ChatGPT

Examples Of Artificial Intelligence
Chatbots Smart Cars
Smart assistants Navigation apps
E-Payments Facial recognition
Search algorithms Text editors
Media streaming Social media feeds

 

Artificial Intelligence Use-Cases
Slide Image Slide Image Slide Image

Healthcare

Companion robots for the care of the elderly

Mining medical records to provide more useful information

Design treatment plans

Assist in repetitive jobs including medication management

Provide consultations

Using avatars in place of patients for clinical training

Heavy Industry

Robots have become common in many industries and
are often given jobs that are considered dangerous to humans.

Robots have proven effective in jobs that are very repetitive which may lead
to mistakes or accidents due to a lapse in concentration and other jobs
which humans may find degrading.

Finance

Algorithmic

Trading

Market Analysis and Data Mining

Personal Finance

Portfolio Management Underwriting

 

 

Artifical Intelligence…
Advantages Disadvantages
  • More powerful and more useful computers.
  • New and improved interfaces.
  • Solving new problems.
  • Better handling of information.
  • Relieves information overload.
  • Conversion of information into knowledge.

 
  • Increased costs
  • Difficulty with software development slow and expensive
  • Few experienced programmers
  • Few practical products have reached the market as yet.

 

Machine Learning Definition
Traditional system development relies on human problem-solving talents and programming skills.

Machine learning algorithms "learn" independently how to tackle a specific problem by utilizing massive volumes of data and substantial trial-and-error, frequently yielding unique insights and superhuman skills.
 

 

The Black-box Problem

Machine learning models, particularly neural networks, have a tendency to become inexplicable to humans, including their developers.

 

AI Model

 

Input

Output

 

What is generative AI?
Generative AI is a type of artificial intelligence that creates new content (text, images, audio, or video) depending on the information we input. 

The term "generative" means that something is created or produced. 
That is exactly what this type of AI does: it generates content that mimics what a human would create.

The most well-known generative AI service right now is ChatGPT, although there are many others. 

They all have one thing in common: you can write to them like you would in a chat and receive responses in text, image, or audio format.
Slide Image

AI technology is not new. You probably already use it in your everyday life, without thinking about it.

For example, when: 
Spotify suggests music you might like, 

  • Google Maps shows you the fastest route, 
  • Your phone sorts photos by people. 
     

What's new about generative AI is that it doesn't just analyse - it creates. And it opens up many new ways to get help in the work!

Slide Image
Slide Image

How does AI generate anything new, and what do you need to know to grasp how it operates in the background? Here's an in-depth look at how to use AI intelligently and safely.

As previously stated, ChatGPT and Copilot use language models that have been trained on massive amounts of text, such as books, papers, and web pages. They learn to identify patterns in how people write and communicate. When you ask a question, the AI attempts to guess a possible response—it has no idea what's correct or wrong.

Consider it the world's most advanced autocomplete. A language model, like your mobile phone, proposes the next word as you enter a message - but on a much wider scale and with a better comprehension of the context.

Keep in mind!
AI doesn't have the understanding of a human.
It lacks consciousness, emotions and its own experiences.

AI predicts responses based on what it has seen most of – not what is true.
It reproduces the patterns it has learned are statistically probable.

AI can sometimes be wrong or even make things up.
These are called hallucinations and occur when the model generates information that does not correspond to reality.

AI does not understand causation or ethics.
It can generate texts on these topics, but has no real insight into why something is happening or what is morally right. After basic training, many models get a second phase: humans review the answers and teach the AI which answers are helpful and harmless. This acts as an imposed safety filter, but it is a learned behavior – not an ethical understanding.

Slide Image
Slide Image

It might not always be clear how generative AI may help with your specific task, but a smart first step is to ask yourself:

What takes up a lot of time in my daily life?

Are there any sections that are repetitious, time-consuming, or difficult to get started on?

Then AI can be useful.

Useful Tip
  Begin by listing 2-3 routine tasks in your business that could benefit from AI, like email drafting, to prepare for hands-on applications ahead.
 
Assess your current tools—do they use AI? 
 
This self-check paves the way for integration strategies in future modules.

 

 

Exploring Key AI Concepts
Machine Learning (ML): 
AI that learns from data without explicit programming to make predictions, e.g., recommending products based on past sales, forecasting inventory for a service firm.
Deep Learning: 
Advanced ML using neural networks for complex tasks like image recognition (e.g., identifying defects in manufacturing).

 

Differentiating AI vs. Automation
 

 

Automation:

Fixed, rule-based processes, e.g., a spreadsheet auto-summing sales in a MSME.

Artificial Intelligence (AI):

Adaptive and learning-based, learns from data to handle variability, e.g., chatbots adapting to customer queries.

Difference:
Automation is rigid, following fixed rules; AI evolves with new data, ideal for uncertain markets (Organisation for Economic Co-operation and Development "Artificial Intelligence Review" 2024).

 

 

 

Practical Example
A craft brewery automates bottling labels (fixed rules) but uses AI to predict flavour trends from customer feedback, adapting recipes seasonally, leading to sales growth and readiness for predictive analytics. 
Assessing AI's Impact on Business Growth
Benefits: Increased efficiency, better decision-making, innovation, and new opportunities (e.g., personalized marketing for a small e-commerce business). Risks: Data dependency—poor data leads to errors; start with clean, local data.
Case Study:
A MSME in logistics used AI for route optimization, cutting costs by 20% (Deloitte "Tech Trends" 2023).
Practical Example: 
An online store implements AI for trend analysis, identifying popular products early and boosting revenue. 
Practical Task - Imagine your MSME!

Instructions: 

  • Use a free AI tool like ChatGPT to brainstorm one business application (e.g., sales forecasting). 
  • Write a prompt: "As an AI expert, suggest how a small European retail shop can use ML for inventory without tech expertise."
  • Refine the output: Add your business details.
  • Reflect: How does this align with growth impacts?

Goal: Build curiosity with realistic expectations.

 


 

 

Will AI take over my job?
AI won't replace you – but it can relieve you. Given demographic changes and rising demands for efficiency, we must deploy smart tools to maintain pace.

As an employee, you still have the judgment, experience, and understanding of the situation. 

For example, AI can provide you with a first draft, proposal, or structure, but it is up to you as the person in charge to decide what is good enough.
Think of Generative Al as a digital colleague!
AI can help you get started when you get stuck.

AI can save time on tasks that would otherwise take a long time.

AI can give you new perspectives or formulations.

Remember: AI can't understand context, emotions, or local conditions – that's your role.

 

1.3 AI Regulations & Compliance
0%
Unit Overview
This unit focuses on the legal and ethical landscape of AI, specifically detailing the implications of key regulations such as the EU AI Act for MSMEs to ensure compliance.
 
The emphasis is on practical, "what you must do" steps, including legal requirements, documentation, and compliance checklists, without deep dives into enforcement mechanisms.

 

Laws, regulations in the use of AI
In this section, we will discuss the general regulations that govern the usage of artificial intelligence in organizations. Slide Image

Why regulate?
Regulations ensure safe AI use; for European MSMEs, compliance avoids fines (up to 4% of revenue under GDPR).
Focus: EU AI Act, GDPR, and practical steps for small businesses.

- 75% of SMEs report regulatory awareness as key to adoption (European Court of Auditors "Special Report" 2024).

 
Regulating The Al
  • Al regulations varied significantly across different countries and regions, with some jurisdictions taking more proactive steps than others.
  • Here are some examples of current Al regulations and initiatives.
Slide Image

European Union (EU)

  • The EU has proposed the Artificial Intelligence Act (AI Act), a comprehensive regulatory framework aimed at governing the development and deployment of Al systems within the EU.
  • The proposed regulation categorizes Al systems based on their risk levels and imposes requirements for transparency, accountability, and human oversight.
  • The General Data Protection Regulation (GDPR), although not specifically focused on Al, establishes data protection principles that apply to Al systems, particularly concerning the processing of personal data.AI Act
Slide Image

United States

  • As of now (October 2025), the United States lacks comprehensive federal regulations specifically targeting Al.
  • However, various federal agencies, including the Federal Trade Commission (FTC) and the National Institute of Standards and Technology (NIST), have issued guidance documents and reports on Al-related topics such as fairness, transparency, and accountability.
  • Some states, such as California, have enacted legislation addressing specific Al-related issues, including privacy (e.g., California Consumer Privacy Act) and automated decision-making (e.g., Automated Decision Systems Accountability Act).
Slide Image

Canada

Canada has established the Pan-Canadian Artificial Intelligence Strategy, a federal initiative aimed at fostering Al research and innovation in Canada.

The strategy includes investments in Al research, talent development, and commercialization efforts, as well as initiatives to address ethical and societal implications of Al.

Slide Image

China

  • China has issued guidelines and standards for Al development, focusing on areas such as Al ethics, data security, and industry standards.
  • The Chinese government has also implemented regulatory measures to govern Al applications in sectors such as finance, healthcare, and transportation.
Slide Image

International Organizations

International organizations such as the Organisation for Economic Co-operation and Development (OECD) and the United Nations (UN) have developed guidelines and principles for Al ethics and governance.

For example, the OECD has issued the OECD Principles on Al, which provide recommendations for responsible Al development and deployment.

Slide Image

 

 

Why the EU got it right: regulating data before AI!

The General Data Protection Regulation (GDPR) is a privacy and security law drafted and passed by the European Union (EU), that imposes obligations onto organizations anywhere.

This goes back to the right to privacy, part of the 1950 European Convention on Human Rights: “Everyone has the right to respect for his private and family life, his home and his correspondence”

Slide Image
Moreover, the EU-GDPR is all about people’s privacy rights:
  • The right to be informed
  • The right of access
  • The right to rectification
  • The right to erasure
  • The right to restrict processing
  • The right to data portability
  • The right to object
  • Rights in relation to automated decision making and profiling.
Slide Image

 

Can we trust AI? Towards “Trustworthy AI” in the EU
Slide Image

The European Commission appointed a group of experts to provide advice on its artificial intelligence strategy: High-Level Expert Group on AI.

According to the Guidelines, trustworthy AI should be:

  • Lawful: respecting all applicable laws and regulations.
  • Ethical: respecting ethical principles and values.
  • Robust: both from a technical perspective while taking into account its social environment.
Several important guidelines were proposed:
  • Human agency and oversight: AI systems should empower human beings.
  • Technical Robustness and safety: AI systems need to be resilient and secure.
  • Privacy and data governance: data governance mechanisms must be ensured.
  • Transparency: the data, system and AI business models should be transparent.
  • Diversity, non-discrimination and fairness: AI systems should be accessible to all.
  • Societal and environmental well-being: AI systems should benefit all human beings.
  • Accountability: ensure responsibility and accountability for AI systems and their outcomes.
Slide Image

 

The AI Act and how it will impact our lives

The AI Act is a document proposed by the European Commission that contains several harmonised rules regarding AI applications;

  • emphasising that its approach is shaped by EU values and risk-based, ensuring both safety and fundamental rights protection.
Slide Image

EU AI Act - Regulation COM/2021/206

Goals:

  1. To provide a single framework for Al products and services that are placed on the EU market, even where providers are not based in the Union;
  2. To ensure that Al systems are safe, secure, trustworthy and respect fundamental rights and values.
  • Definition of an Al system: aligned with the OECD; “machine-based system designed to operate with autonomy ... infers, from input received, how to generate output...
  • "Scope: applicability for providers, deployers, importers, distributors and manufacturers of Al systems across all EU Member States, with extraterritorial effect similar to the GDPR
  • Regulatory approach: modelled on EU product safety legislation, with a risk-based approach
    • Obligations ∞ level of risk
Slide Image

What does the AI Act propose?

  • Prohibition of unacceptable AI practices (e.g., social scoring)
  • Regulation of high-risk AI systems (e.g., AI used in the context of recruitment)
  • Conformity assessment (i.e., under the EU product safety framework)
  • Transparency obligations for potentially deceptive AI systems
  • Ex post market surveillance (i.e., post-market monitoring system)
  • Governance (i.e., authorities must be appointed for the application and implementation)
  • Pre-emption of national AI regulatory frameworks (i.e., regulated by the EU)
  • Monitoring and enforcement (i.e., done by the Member States)
  • Compliance with the prohibitions and regulatory requirements
Slide Image

 

EU AI Act – The tiered, risk-based approach

Slide Image

EU AI Act - Prohibited Al systems & High Risk Al Systems

Slide Image

Unacceptable risk - Al systems that pose a clear threat to the health, safety and rights of people will be prohibited.

  • Examples: Al systems deploying subliminal, deceptive techniques with the objective or effect of impairing a person's ability to make an informed decision; biometric categorization systems; social scoring by governments; real-time biometric identification in public spaces.
 
Slide Image

High risk - Al systems considered to pose a high risk to the health, safety, and rights of people will be subject to strict obligations.

  • Examples: Al-powered critical infrastructure; the use of Al in the safety components of products, or the provision of essential public and private services; employment; education; administration of justice. Full list in Annex III.
 
  Note: citizens will have the right to file complaints w/designated authorities  

 

EU AI Act - Obligations for High Risk Al Systems

  • Conformity Assessment
    • Risk identification and mitigation exercise; transparency and accountability; human oversight; data quality.
  • Fundamental Rights Impact Assessment
    • For deployers of high-risk Al that are bodies governed by public law or private operators providing public services.
    • Intended purpose of the Al system; duration and frequency of deployment; individuals or groups likely to be affected; specific risks of harm; measures to be taken to mitigate such harms (including governance and complaint mechanisms).

Note: Symbiosis between the different actors identified within the Al Act.

Slide Image

EU AI Act - A Note on General Purpose Al Models / Systemic Risk

  • GPAI models fall under Article 52 of the Al Act;
  • All providers of GPAI are subject to certain conditions including:
    • Technical documentation; compliance with EU copyright law;
    • Providing information to Al system providers who intend to use the GPAI model;
  • Additional obligations for providers of GPAI models that pose systemic risk:
    • Model evaluation; Risk assessment and incident reporting; cybersecurity protections;
    • Labeling obligations
    • Models may be designated as presenting systemic risk ex officio by the Commission, if they meet a computation training threshold, or if the model in question has "high impact capabilities.

"Note: Obligations above do not apply to any Al systems or GPAI models where they are specifically developed for scientific research and development (Article 2(5a)).

How the AI Act Relates to Other Regulations:

Interconnections:
Complements GDPR (data handling in AI), NIS2 Directive (cybersecurity), product liability laws.
Harmonization:
Think of avoiding overlaps through unified compliance strategies. àUnified strategies—e.g., one DPIA covers both.
What You Must Do:
Create a compliance matrix linking regulations; assign roles and responsibilities (European Commission "AI Act" 2024).
 
   

 

EU AI Act & Interplay with GDPR

The GDPR continues to apply to personal data processing:

The Al Act:

  • Should not be understood to indicate compliance with other EU law, including the GDPR (Rec. 27).
  • Does not guarantee compliance with Article 9 "special category" rules established under the GDPR (Rec. 24).
  • Relies on definitions of "biometric data" established in the GDPR (Rec. 7).
  • Commission launches "Code of Practice" for general purpose Al technologies (July 2024).

European Data Protection Authorities (DPAs) have remained active in the Al space:

  • Ongoing Italian DPA case against OpenAl's ChatGPT;
  • Netherlands fines Clearview for FR data collection
  • DSK publishes "Hambach Declaration on Al" - collective guidance document for Germany

How the AI Act Relates to Other Regulations:

MSME Benefit: Streamlines for small teams, avoiding overlaps.

Practical Example: A MSME creates a matrix linking AI Act and GDPR for e-commerce tools, simplifying audits and enabling secure personalization.

Useful Matrix Example: Columns: Regulation, Requirements, MSME Actions (e.g., GDPR: Consent; Action: Forms).

Slide Image
 
Summary
Slide Image
  • The development of data-driven artificial intelligence applications is pushing the limits of the applications of these algorithms in our lives: this rapid evolution motivated the need to ethical, legal and technical regulatory frameworks based on specific principles: accountability, interpretability, fairness, safety, privacy.
  • At the European level, there have been several proposals to regulate data and AI-based applications: EU-GDPR, Trustworthy AI Initiative and AI Act.
  • Open regulatory challenges will focus on the impacts of AI in ethics, transparency, fairness, safety, sociology and sustainability.
  • Multidisciplinary work is, more than ever, of utmost importance and useful.

 

How to Comply…

How to Comply…

 

Overview of the EU AI Act

Structure: Risk-based (prohibited: e.g., social scoring; high-risk: e.g., AI hiring; limited/minimal: e.g., chatbots).

MSME Implications: Classify your AI (most MSMEs use minimal-risk); require transparency for limited-risk.

What You Must Do: Conduct risk assessment checklist; maintain usage logs (European Commission "AI Act" 2024).

Example: A MSME using AI for customer support must label it as AI-generated.

This risk framework is crucial for ethical deployments, like chatbots for service.

Practical Example: A retail MSME classifies its inventory AI as minimal-risk, logging usage to comply—saving time for growth-focused toolz.

 

Navigating Data Privacy Laws (e.g., GDPR):

Principles: Consent, minimization (only collect needed data), lawful processing.

What You Must Do: DPIA for high-risk AI; get explicit consent (European Data Protection Board "Opinion" 2024).

MSME Example: A service MSME uses GDPR-compliant AI for email personalization, boosting engagement 18%.

Practical Example: A bakery MSME conducts DPIA for AI customer loyalty apps, ensuring data minimization and gaining 12% repeat business—preparing for ethical bias checks in Unit 1.3.

 

Addressing Legal Considerations for AI Deployment:

 

Liability: MSMEs liable for AI harms (e.g., wrong recommendations); use vendor contracts with clauses.

What You Must Do: Review tools pre-deployment; document decisions (Organisation for Economic Co-operation and Development "Governing with AI" 2025).

MSME Tip: For a small manufacturer, audit AI suppliers for compliance.

Practical Example: A consulting MSME includes liability clauses in AI software contracts for client reports, preventing disputes and enabling expansion to customer service bots.

Complying with Industry-Specific AI Regulations:

Examples: Healthcare (medical devices regulations); Finance (anti-money laundering rules); Retail (consumer protection laws).

MSME Focus: Check sector guidelines—e.g., retail MSMEs ensure AI ads comply with fairness laws.

What You Must Do: Identify applicable sector rules and integrate them into AI planning. Use checklists; consult free EU resources.

Practical Example: A finance MSME complies with AML for AI fraud detection, reducing risks by 30%.

 

Implementing ISO 42001:

Overview:

International standard for AI management systems, focusing on governance and risk management.

What You Must Do: Establish an AI policy; audit annually; train staff on basics (International Organization for Standardization "ISO/IEC 42001" 2023).

Example: A MSME implements for AI inventory, gaining client trust. Find it in our AI digital repository!

Practical Example: A design MSME drafts a simple AI policy for tool use, auditing quarterly to certify; boosting partnerships and readiness for Module 4's loyalty programs.

 

Practical Task:

Instructions:

  • For your MSME, classify a potential AI tool (e.g., chatbot) under EU AI Act.
  • Prompt an AI: "Assess risk level for AI customer support in a small European shop per EU AI Act."
  • Add checks: Include GDPR consent steps.
  • Refine: Propose a compliance matrix.

Goal: Prioritize legal attitudes.

1.4 Ethical and Responsible AI Use
0%
Unit Overview
AI can be powerful, but it must be used fairly and safely.
This unit explains how MSMEs can avoid unfair bias, protect sensitive data, and remain transparent with staff and customers. 
The key message is that AI should always support people, not replace their judgment.
The unit includes practical tips, checklists, and MSME scenarios (e.g., using AI in employee scheduling without discrimination).
It builds on regulations from Unit 1.2 but focuses on voluntary best practices.

What do we mean by “ethics”?

Our understanding of what is right and wrong, as well as the rules that individuals and the society should follow.

What do we mean by “AI ethics”?

AI ethics refers to the moral principles and values that should guide the development and use of artificial intelligence systems.
AI systems should be designed and used in a way that benefits society and follows ethical principles.
Issues like fairness, transparency, privacy and safety are important ethical considerations.

Slide Image
Why does ethics matter for AI?

AI has the potential for immense impact, both positive and negative.

AI systems are increasingly being entrusted with high-stakes decisions that affect human lives and society.

AI learns from data, which can encode human biases and prejudices. Unethical applications could automate or worsen injustice.

As AI gets more capable and autonomous, we need to maintain human accountability and control.

Ethical issues around AI use, like privacy invasion or job losses, require careful consideration to avoid public backlash.

Ethics helps ensure AI development and use is guided in a socially responsible direction.

  • Responsible AI is a framework that guides how we should address the challenges around artificial intelligence from both an ethical, technical and legal point of view.
    • We must resolve ambiguity for where responsibility lies if something goes wrong!
  • This framework relies on fundamental principles:
    • Accountability
    • Interpretability
    • Fairness
    • Safety
    • Privacy

 

Accountability: who will take the responsibility?
  • People should be accountable for AI systems.
    • This principle is the baseline, hence, all the other principles can be seen as branches.
  • Some important ideas are:
    • Implement and use a human-centered design approach.
    • Identify multiple metrics to assess training and monitoring to ensure that the metrics are appropriate for the context and goals of system.
    • Examine the raw data (e.g, missing values, incorrect labels, biases, feature redundancy).
    • Understand the limitations of databases and models.
    • Learn best to make sure the AI system is working as intended and can be trusted.
    • Continue to test, monitor and update the system after deployment.
Fairness: how to deal with bias?
  • AI models learn from existing data collected from the real world, and so an accurate model may learn or even amplify problematic pre-existing biases in the data based on sensitive characteristics
  • Regarding this issue, we should:
    • Interact with social scientists, and other relevant experts to understand and account for various perspectives
    • Consider how the technology and its development over time will impact different use cases (e.g., what outcomes does this technology enable)
    • Assess fairness in our datasets (e.g., identifying representation and corresponding limitations)
    • Check the system for unfair biases
    • Analyse the performance of the system, taking into account the different metrics.
Slide Image
Safety: achieving reliable and safe AI systems
  • Safety and security ensure that AI systems behave as intended, regardless of how attackers try to interfere.
  • Regarding this issue, we should:
    • Consider if there are incentives to make the system misbehave.
    • Identify what unintended consequences would result from the system making a mistake and assess the likelihood and severity of these consequences.
    • Build a rigorous threat model to understand all possible attack vectors.
    • Research into adversarial machine learning, as it continues to offer improved performance for defenses and provable guarantees.
    • Check if there are other vulnerabilities in the AI supply chain.
Slide Image
Privacy: can AI reveal aspects of its training data?
  • AI systems must prioritise and safeguard consumers’ privacy and data rights and provide explicit assurances to users about how their personal data will be used and protected.
  • Regarding this issue, we should:
    • Identify whether AI models can be trained without the use of sensitive data.
    • Anonymise and aggregate incoming data using best practice data-scrubbing pipelines (e.g., removing personally identifiable information (PII) and outlier or metadata values that might allow de-anonymisation).
    • Train models using federated learning, where a fleet of devices coordinates to train a shared global model from locally-stored training data.
    • Perform tests based on “exposure” measurements or membership inference assessment to estimate whether the model is unintentionally memorising or exposing sensitive data.

 

Ethical Considerations for Al Development
Human Agency and Oversight: Technical Robustness and Safety:  Privacy and Data Governance:
All decisions are reviewed by humans and humans are the "in charge" of final decisions. Al tools do not contain errors and their operation does not lead to harm. There are strict rules about people's data and how Al can use it.
Diversity, Non-Discrimination and Fairness
Al should work equally well for differen people and in different contexts. The benefits of Al should be shared equally.
Societal and Environmental Well Being
Al should impact the environment and the society in a positive way.
Transparency
It is possible to say what decisions Al makes, what steps led it to those decisions and what data informed them.
Accountability
We know who is responsible for the Al and we can fix problems or enforce sanctions when something goes wrong
Will we all be able to find work in a world with AI?

Automation is often seen as a threat to manual labour. Robots replace factory workers and farmers, while self-driving cars come after human drivers' jobs.

But Al is also trained to perform more intellectual tasks such as translation, programming, writing, medical diagnosis or teaching.

Slide Image

 

Questions to consider
  • Who should be blamed when decisions made by Al lead to harm?
  • What are the differences between the rules that should apply to ethical Al and those that govern our own ethical behaviour?
  • To what extent can we ignore the risks of Al to enjoy the benefits it provides?
  • What can we do as citizens to influence how Al is used?

 

What ethical problems can AI pose?

From algorithmic bias and privacy violations to questions of accountability and autonomy, the ethical dimensions of Al are increasingly profound and multifaceted.

The following delves into the intricate web of ethical issues surrounding artificial intelligence, shedding light on the crucial considerations that underpin the responsible development and deployment of Al technologies.

Slide Image

Ethical problems AI can pose

Bias and Fairness:
Al systems can inherit biases present in their training data, leading to discrimination against certain groups. Ensuring fairness in Al decision-making is a significant challenge.

Privacy: 
Al technologies, especially in surveillance and data analysis, can infringe upon individuals' privacy rights. Striking a balance between security and personal privacy is crucial.

Accountability:
Determining who is responsible when Al systems make errors or harmful decisions is often unclear. Establishing legal and ethical frameworks for accountability is a challenge.

Transparency: 
The inner workings of some Al algorithms are opaque, making it difficult to understand why they make certain decisions. This lack of transparency can undermine trust.

More ethical problems that AI can pose

Security Risks:
The use of Al in cyberattacks and the potential for Al to generate misinformation pose security risks. Safeguarding against these threats is an ongoing challenge.

Ethics in Al Research: 
Ethical considerations extend to the very research and development of Al, including issues related to data collection, experimentation, and collaboration.

Ethical Alignment:
Al researchers and engineers often face dilemmas about whether to align their work with ethical principles or prioritize commercial interests. Striking a balance can be challenging.

Intellectual Property and Open Source: 
The debate about open source versus proprietary Al technology raises ethical questions about access to and control over Al innovations.

How would you use AI?

Ethical Al use involves transparent data handling, minimizing bias, ensuring privacy, and making decisions in the best interest of society.

It requires ongoing vigilance, open communication, and a commitment to responsible Al development to maximize benefits while mitigating harm.

Slide Image

 

Potential ethical gains of AI Effectivity

Effectivity: Automation, precision, speed and scalability of processes implies lower expenditure and a better prison service overall.

Prevention: AI-systems capable of detecting undue behavior and alerting officers may lower criminality, suicides, violence and smuggling of contraband.

Prediction: The power of machine learning may be harnessed to generate novel insights and predictions which may be used to decrease recidivism.

Intelligence: Data-mining may lead to knowledge that can be shared with other governmental bodies or used in united action with the police.

Justice: Humans are notoriously bad at making fair decisions. We are often prejudiced, inconsistent and often make assessments that are affected by our personal feelings. A well designed AI-system may well lead to greater justice in decision-making processes.

Slide Image

 

Mitigating Risk
Data quality: Good quality data that is representative and generated with awareness of ethical risks will mitigate the tendency of models to become biased or skewed.
Knowledge & awareness: Mitigating risk starts with being aware of the risks and knowing in what situations they are likely to occur. Education for all organizational members may be required as well as continuous research among specialists.
Governance: Management processes and governance structures that align AI-solutions with strategic objectives of the organization. Enhancing performance, mitigating risk and assuring the best interest of all stakeholders.
Center of excellence: An interdisciplinary team dedicated to maintenance, development and continuous improvement of AI-solutions in the organization, delivering technical functionality and novel insights directed towards achieving organizational goals.
Transnational cooperation: Sharing data, insights and best practices across borders will promote effective solutions and wiser implementation of advanced AI in prison service

 

The starting point in your work with AI should always be to see AI as a support, not a decision-maker. This means that you are responsible for what you deliver or produce, even if an AI has given you suggestions or documentation. AI is not a substitute for human judgment.

On several occasions, lawyers have tried to ask, for example, Chat GPT for legal support for certain questions, but have discovered in the control of the answers that they do not at all comply with the current legal framework.

So, the answers that AI gives you don't have to be right, even if the language it uses sounds correct. Therefore, it is always important to be “source-critical” and double-check the answers you have received from AI!

Reflect: When AI becomes part of the decision/work process - who is ultimately responsible?
What are you allowed to do - and what should you NOT do?

In this part, we'll talk about what you get and what you should leave out when it comes to working with AI.

More and more people want to use AI services such as Copilot, ChatGPT or other text-based tools in their work.

This is positive – but it is important to understand what you can and cannot do.

 

The basic rule is simple:

You may NOT paste confidential information, personal data, or sensitive information into external AI services.

This is true even if it feels like "just a draft" or if you use an AI service via a web browser on your work computer. 

Why?
AI services are often located in other countries, and we don't know exactly how data is handled or stored, for example.

The information can leave our control – and it must not happen!
 

 

What is meant by external AI services?

External AI services are those that are delivered by external actors and that your company itself does not manage or have full control over. 
Examples are Chat Tools on the web or AI services from commercial platforms.

These are often located in other countries, and we don't know exactly how data is handled.

And in-house AI solutions?
These are AI systems that the company itself is responsible for, or that are delivered via the company’s own IT environments where you have control over data protection and storage.

 

Reflect:
How can companies benefit from AI services in general without compromising legal requirements, information security or the municipality's control over its own material?

 

AI supports – YOU CONTROL!

You won't be replaced by AI.

It is you, who know your business, your colleagues and your needs – who create real benefits with AI.


Technology is a tool, but it is in combination with your experience and understanding that it becomes a powerful tool.


When used correctly, there is great potential to be able to streamline your work and free up time for more value-creating tasks.

 

Example
An example could be as support for staff in the care sector. 
  • Here, AI can, for example, make it easier for staff to find information such as checklists and routines. 
  • AI can also simplify the texts or translate them into another language. This gives staff more space to be close to the user, and at the same time reduces the risk of deviations. 
  • Working with AI is also a great opportunity to explore new areas and learn more. AI can provide inspiration, new perspectives and help you quickly get an overview of topics you are curious about. 
  • So use AI both as a support where you are already safe – and as a tool to grow. The best of both worlds!

 

Case Studies

Case study 1: AI in Healthcare: 

The integration of Artificial Intelligence (AI) into healthcare has brought about a transformative shift in how medical diagnosis, treatment, and patient care are approached. However, while Al offers numerous benefits, its use in these critical areas also raises significant ethical concerns that must be carefully considered to ensure that technology serves the best interests of all patients.
 

Case study 2: AI in Autonomous Vehicles:

 Autonomous vehicles (AVs) represent a ground-breaking innovation in transportation, promising to reduce traffic accidents, improve road safety, and revolutionize mobility. However, the deployment of AVS also introduces profound ethical dilemmas, particularly in scenarios where the vehicle must make decisions involving life-and-death situations. Unlike human drivers.
 

 

How do we avoid Algorithmic Bias?

Diversity is the key.

1. Diversity of Background of the Team (to avoid "group think")

2. Diversity of Mindset (personality testing)

3. Diversity of Data

4. Diversity of Algorithmic Models

Slide Image

 

Protecting Privacy and Data Security

Principles: Anonymize data where possible and limit access.

MSME Examples: Encrypt and secure customer data in AI tools, such as with the use of AI chatbots.

Safeguards: Implement basic encryption and use consent forms.

Keep in mind: Secure data practices enable predictive analytics without breaches.

Building Transparency and Accountability

Communication:

Inform employees and customers about AI use (e.g., labels like "AI-generated response").

Documentation:

Keep logs of AI decisions for audits.

Accountability:

Assign roles for AI oversight in the business.

Practical Example: A MSME assigns an "AI overseer" role for tool logs, ensuring accountability and reducing errors.
 
Useful Log Template: Date, AI Use, Decision, Review Notes.
Promoting Human-Centered AI

Role of Humans: AI as a tool, not replacement—e.g., using AI for routine tasks to free up time for creativity.

Ethical Decision-Making: Frameworks for when to override AI (e.g., in ethical dilemmas).

Fostering Trust: Training staff on AI ethics to build internal confidence.

MSME: A shop uses AI for stock but humans for decisions.

Practical Example: A cafe MSME trains staff to override AI suggestions in customer disputes, maintaining relationships and boosting satisfaction.

Practical Task:

Instructions:

  • Brainstorm AI bias in your MSME (e.g., customer targeting).
  • Describe issue (5-10 sentences).
  • Propose 3 mitigations with human oversight.
  • Example: "AI ads target only young demographics—mitigate by diverse data."

Goal: Value accountability.

Practical Example: A MSME describes hiring AI bias, mitigates with diverse training data and human reviews; estimating fairer outcomes.

 

 

Summary
0%
Summary

 

AI is Augmentation

It is a strategic tool that relieves teams from repetitive tasks, driving efficiency and innovation in MSMEs

 

Ethics is Governance

Principles like Fairness, Transparency, and Accountability are not optional; they are the core of legal compliance and trustworthy market engagement.

   

Regulation is Risk-Based

The EU framework (GDPR and AI Act) requires proportional compliance. MSMEs must classify their AI tools' risk level.

 

Key to Success

Start small, prioritize ethics! Remember: AI augments, not replaces!

Identify one low-risk, high-value and automate it with an AI tool to test the water.

Self-assessment test
  • Provider: CCG EQF Level: 4
Keywords: Artificial IntelligenceAIAI FundamentalsMSMEsEU AI ActGDPRIOS42001Ethical AIMachine LearningDeep LearningAI ComplianceResponsible AIBias MitigationData Privacy
Objectives and Learning Outcomes
Objectives:

●    To provide MSME owners, managers, and staff with a foundational understanding of Artificial Intelligence (AI) concepts, tailored to business applications without requiring technical expertise. 
●    To outline key regulatory requirements, such as the EU AI Act and GDPR, and their practical implications for compliance in small-scale operations. 
●    To foster awareness of ethical considerations in AI use, emphasizing fairness, transparency, and human-centered approaches to build trust and sustainability. 
●    To equip participants with practical takeaways for integrating AI responsibly into daily business processes, such as customer service or inventory management.
 

Learning Outcomes:

Knowledge: 
•    Understand basic AI definitions and concepts, including how AI differs from traditional tools; 
•    Comprehend basic AI regulatory frameworks and their relevance to business operations; 
•    Understand key principles of fairness, transparency, and privacy in AI use.
Skills: 
•    Articulate AI's potential business value through simple examples; 
•    Identify common compliance requirements and apply basic checks (e.g., risk assessments); 
•    Identify responsible practices and apply simple safeguards (e.g., reviewing outputs, protecting sensitive data). 
Attitudes: 
•    Develop curiosity towards AI's capabilities while maintaining realistic expectations; 
•    Prioritize ethical and legal considerations in AI adoption to build trust and avoid penalties; 
•    Value accountability and maintain trust in how AI is applied. 

Suggested Prompt

For Trainers (AI-Assisted Session Preparation): 
-    "As an AI expert, generate a 5-slide PPT outline for introducing AI fundamentals to MSME owners, including one real-world example from retail, one analogy for machine learning, and a myth-busting bullet point. Ensure language is simple and engaging for non-technical audiences."
-    Create 5 interactive discussion questions based on AI myths vs. realities for MSMEs, incorporating real-world examples from retail or services to foster curiosity and realistic expectations about AI adoption.”

For Learners (Self-Reflection): 
-    "Reflect on your MSME: Describe one way you could use AI (e.g., for inventory management) while ensuring compliance with the EU AI Act's risk classification—include a simple checklist of 'what you must do' steps."
-    "As an AI assistant, help me develop a simple compliance checklist for implementing AI in my small retail business. Focus on EU AI Act requirements, GDPR data handling, and ethical considerations like bias mitigation and transparency. Include step-by-step actions tailored for MSMEs, using examples like customer recommendation systems or inventory forecasting. Ensure the response is practical, non-technical, and emphasizes human oversight."

For Ethical Discussion: 
-    "Using the principles from Unit 1.3, brainstorm a scenario where AI bias could affect your business (e.g., customer targeting) and outline 3 mitigation steps, emphasizing human oversight."

For self-study: 
-    "Explain how ISO 42001 can help a small service provider establish an AI policy, including benefits for compliance and risk management, in simple terms with a sample policy outline."
For assessment: 
-    "Design a multiple-choice quiz on ethical AI principles, covering fairness, transparency, and privacy, with scenarios relevant to MSMEs like using AI chatbots for customer service."
 

Glossary
Artificial Intelligence (AI)

Systems or machines that mimic human intelligence to perform tasks and can iteratively improve themselves based on the information they collect

Machine Learning (ML)

A subset of AI that enables computers to learn and improve from experience without being explicitly programmed, often using data to make predictions or decisions.

EU AI Act

A comprehensive European Union regulation (effective from 2024, with phased implementation) that categorizes AI systems by risk levels and imposes requirements for safety, transparency, and accountability.

GDPR (General Data Protection Regulation)

EU law on data protection and privacy that addresses the export of personal data outside the EU and requires consent and lawful processing of personal data.

Algorithmic Bias

Systematic and repeatable errors in a computer system that create unfair outcomes, often due to prejudiced assumptions in the machine learning process.

References

“Artificial Intelligence Adoption by SMEs to Achieve Sustainable Business Performance.” Sustainability, vol. 16, no. 5, 2024, https://www.mdpi.com/2071-1050/16/5/1864.

×
Asistente de lectura