public:t-709-aies-2025:aies-2025:principle_ai_ethics
Differences
This shows you the differences between two versions of the page.
Next revision | Previous revision | ||
public:t-709-aies-2025:aies-2025:principle_ai_ethics [2025/09/18 08:34] – created leonard | public:t-709-aies-2025:aies-2025:principle_ai_ethics [2025/09/22 12:41] (current) – leonard | ||
---|---|---|---|
Line 16: | Line 16: | ||
* Responsibility | * Responsibility | ||
* Privacy | * Privacy | ||
+ | |||
+ | Taken from: Jobin, et al. "The global landscape of AI ethics guidelines." | ||
+ | |||
+ | UNESCO added principle (taken from: UNESCO. “Recommendations on the ethics of Artificial Intelligence" | ||
+ | |||
+ | * Proportionality | ||
+ | |||
=== Transparency === | === Transparency === | ||
| What is it? | Understandability of decisions/ operations/ plans of systems | | | What is it? | Understandability of decisions/ operations/ plans of systems | | ||
- | | Importance | + | | Importance | Building trust, scrutinizing the issues, and minimizing harm | |
| Issues | | Issues | ||
| Implementation | | Implementation | ||
+ | |||
+ | * If a public sector or institution makes a decision, the people whose rights have been affected should know whether an AI system plays a role in decision-making. | ||
+ | * In such cases, the people have the right to request explanations of why and how the decision has been made. The institution should provide information and revise the decision if proper explanations do not exist. | ||
+ | * The designers whose AI systems affect the rights of other humans should commit to choosing/ | ||
=== Justice and Fairness === | === Justice and Fairness === | ||
Line 52: | Line 63: | ||
| Implementaiton | | Implementaiton | ||
+ | === Proportionality === | ||
+ | |||
+ | | What is it? | - AI systems must not be used beyond what is necessary to achieve a legitimate aim.\\ - Developers and deployers should assess and prevent harms, requiring that any use of AI be appropriately scaled and carefully considered relative to its purpose | | ||
+ | | Importance | ||
+ | | Issues | ||
+ | | Implementation | ||
+ | |||
+ | |||
+ | ===== AI - What are we talking about? ===== | ||
+ | |||
+ | ==== Historical Perspective ==== | ||
+ | |||
+ | In 1950, Turing proposed the idea of building child machines, which are machines that can mimic human children' | ||
+ | |||
+ | In 1956, McCarthy and Minsky proposed ideas for building reasoning systems based on logic rules and symbolic representations. In the 1960s and 1970s, heavy emphasis was put on symbolic AI and rule-based systems such as chess. | ||
+ | |||
+ | The 1970s and 1980s marked the AI winter, where the progress in AI research slowed down. | ||
+ | |||
+ | In the 1990s and 2000s, AI matured in the form of machine learning (ML) algorithms, which are data-driven methods, as opposed to rule-based symbolic systems. | ||
+ | |||
+ | |||
+ | ==== Contemporary AI: Practical Artificial Neural Networks ==== | ||
+ | |||
+ | In the 2010s, Artificial Neural Networks (ANNs) required enormous amounts of data and computing power. ANN led to extensive automation in the domains of | ||
+ | * Image recognition: | ||
+ | * Natural Language Processing: text and audio data | ||
+ | |||
+ | In recent years, ANNs have been used in building large language models (LLMs), e.g., GPT models. | ||
+ | * Training data: texts on the internet. | ||
+ | * How do ANNs get training? | ||
+ | * Training: feed input data (examples) with known outputs, which tunes the weights (edges between nodes) until it predicts well. | ||
+ | * Backpropagation: | ||
+ | |||
+ | |||
+ | ==== Ethical Issues of ANNs ==== | ||
+ | |||
+ | | Transparency | ||
+ | | Justice and Fairness | ||
+ | | Safety and Security | ||
+ | | Responsibility | ||
+ | | Privacy | ||
+ | |||
+ | |||
+ | ==== Can ANNs become more ethical? ==== | ||
+ | | Explainable ANNs | Using explainability techniques to help users and developers understand the decision. But: Only to some extend. | | ||
+ | | Bias mitigation | ||
+ | | Safety improvement | ||
+ | | Responsibility | ||
+ | | Privacy | ||
+ | |||
+ | |||
+ | ==== Example ==== | ||
+ | |||
+ | === ChatGPT === | ||
+ | * Generative Pre-trained Transformer (GPT) is one of the largest LMMs. | ||
+ | * GPT-4 had 45 TB, GPT-5 even as much as 280 TB of (unfiltered) training data. | ||
+ | * GPT-4 roughly 1.7 trillion parameters. GPT-5 is estimated to have up to 5-10 trillion parameters (could be less than GPT-4, however, due to a possibly multi-model architecture. This is all unclear.). | ||
+ | * Up to 400,000 tokens context window. | ||
+ | |||
+ | **Ethical issues**: | ||
+ | * Transparency: | ||
+ | * Justice and fairness: Use of ChatGPT may violate copyrights, etc. | ||
+ | * Safety: Generating harmful content, misinformation or being misused for unethical purposes. | ||
+ | * Responsibility: | ||
+ | * Privacy: Massive amounts of data. Some training data can be extracted. See for example [[https:// |
/var/www/cadia.ru.is/wiki/data/attic/public/t-709-aies-2025/aies-2025/principle_ai_ethics.1758184491.txt.gz · Last modified: 2025/09/18 08:34 by leonard