Here are your guides with simple words and short sentences:
This activity helps you understand how AI systems can fail and how to make them safer. You will analyze a real AI system and create two outputs:
Choose one of these real-world systems:
For this activity we will use Disaster Relief AI as our example.
Think about how the AI system could fail or cause harm. Write down risks in three areas:
Reliability Risks (These are mistakes the AI makes regularly):
Safety Risks (These are ways people could get hurt physically, financially, or emotionally):
Misuse Risks (These are ways people could use the system unfairly):
Use this two-part framework to understand the system:
Part A: Level of Automation (Where does this system fall?)
Example: Disaster AI is semi-automated because the algorithm suggests which areas need help first and humans review the recommendations.
Part B: Context Criticality (How serious are mistakes?)
Example: Disaster AI has high criticality because wrong predictions can cost lives.
List the parts of the system that could break:
For each failure point ask:
For each part of the system check four trust qualities:
1. Human Control (Can people change the AI's decision?)
Example for Disaster AI: Disaster managers CAN override the AI and send help to different areas if they think the algorithm is wrong.
2. Transparency (Can people understand why the AI made a decision?)
Example for Disaster AI: People can see the weather data and algorithm steps. The neural network calculations are hard to explain though.
3. Fallback Mechanisms (What happens if the AI fails?)
Example for Disaster AI: If the AI system goes down disaster managers can use traditional weather forecasts and manual analysis to make decisions.
4. Accountability (Who is responsible if something goes wrong?)
Example for Disaster AI: The government agency is responsible. They must report errors and fix them.
This is a table you fill in Google Sheets. Here is the template:
| System Part | Automation Level | Human Control | Transparency | Fallback | Accountability | Risk | Safeguards |
|---|---|---|---|---|---|---|---|
| Data Collection | Semi-automated | Low | Low | Manual review | IT + Manager | Medium | Let people check their data |
| Risk Prediction Algorithm | Fully automated | Medium | Low | Traditional forecast | Government agency | High | Add human review step |
| Alert Distribution | Semi-automated | High | Medium | Manual alerts | Communication team | Low | Test system weekly |
For each row:
This is a document in Google Docs that explains everything. Structure it like this:
Part 1: System Overview
Part 2: Key Risks Found
Part 3: Trustworthiness Findings
Part 4: Proposed Safeguards
Part 5: Making It Human-Centered
What is it? An AI system that predicts typhoons, floods and earthquakes in the Philippines. It analyzes weather data, historical events and geography to forecast disasters. It also recommends which areas need evacuation and relief first.
Why does it matter? The Philippines gets about 20 typhoons per year. The system helps save lives by giving early warnings. It also helps decide where to send limited rescue resources and supplies.
Who uses it?
Reliability Risks (System Makes Mistakes)
| Risk | What Could Go Wrong | Real Example |
|---|---|---|
| Bad Data Input | Weather stations in remote areas don't send data | Mountain villages are not tracked and typhoon predictions miss them |
| Model Bias | The AI learned from old data only | System predicts fewer floods in newer flood-prone areas built recently |
| Prediction Errors | Algorithm calculates wrong forecast | AI says typhoon path is safe but it actually goes through a city |
| Outdated Model | System uses old weather patterns | Climate change makes old patterns wrong but system doesn't adapt |
Safety Risks (People Can Get Hurt)
| Risk | Who Gets Hurt | Example |
|---|---|---|
| False Alarms | Entire communities | AI predicts big typhoon but it misses. People evacuate unnecessarily and lose work and income. Trust in warnings drops |
| Missed Warnings | Vulnerable populations | System fails to predict flood. Remote village doesn't evacuate. People drown |
| Wrong Evacuation Orders | Everyone in wrong area | Algorithm says evacuate North but typhoon comes from East. People evacuate to danger |
| Bad Resource Allocation | Communities in real danger | AI sends all rescue teams to Area A but Area B has more people needing help. Area B doesn't get rescued in time |
Misuse Risks (Unfair or Biased Treatment)
| Risk | How It Happens | Impact |
|---|---|---|
| Geographic Bias | System has more data from rich areas | Poor rural areas get fewer warnings and less rescue help |
| Social Bias | Training data reflects historical discrimination | Certain communities are deprioritized for resources |
| Political Pressure | Officials manipulate the system | One region gets resources while another ignored area floods |
| Information Inequality | Only educated people understand warnings | Elderly or non-English speakers miss critical alerts |
Automation Level: Semi-Automated
What this means:
Why this level? The system is not fully automated because disaster decisions are too important to leave to the computer alone. A human disaster manager sees the AI's recommendation and thinks about local knowledge before saying yes or no.
Example:
AI System → Recommends evacuate 50,000 people from Zone A
↓
Disaster Manager → Reviews the recommendation
↓
Manager → Checks if roads are open and if shelters have space
↓
Manager → Approves evacuation OR changes to Zone B insteadContext Criticality: HIGH
What this means:
Why high criticality? Disaster predictions affect life-and-death decisions. A small error in the algorithm can cost lives. When you're dealing with typhoons and floods there is no such thing as a "low stakes" mistake.
1. Data Collection & Input
What could break:
Severity: Medium-High Likelihood: Medium (infrastructure in PH is sometimes unreliable) Who gets hurt: Remote communities that can't be tracked
2. Algorithm & Processing
What could break:
Severity: High Likelihood: Medium (happens with all AI systems) Who gets hurt: Everyone affected by wrong predictions
3. Decision Output
What could break:
Severity: Medium-High Likelihood: Medium (communication can fail) Who gets hurt: Communities that don't understand the warning
4. Human Review & Action
What could break:
Severity: High Likelihood: High (humans are not perfect) Who gets hurt: Communities affected by bad human decisions
5. Feedback & Learning
What could break:
Severity: Medium Likelihood: High (no feedback loop mentioned) Who gets hurt: Everyone because problems never get fixed
| System Component | Automation Level | Human Control | Transparency | Fallback Mechanism | Accountability | Risk Level | Suggested Safeguards |
|---|---|---|---|---|---|---|---|
| Data Collection & Sensors | Semi-automated | Low - sensors collect automatically | Low - citizens don't see raw data | Manual data verification when requested | PAGASA meteorologists | Medium | Allow community check-ups on local stations. Publish data quality reports monthly |
| Risk Prediction Algorithm | Fully automated | Medium - humans review before action | Low - neural network is "black box" | Traditional weather forecast as backup | PAGASA agency | High | Add explainability tool. Show which data points led to prediction. Require human signoff |
| Alert Distribution | Semi-automated | High - humans write final message | Medium - alerts are public | Manual phone calls to mayors if system fails | Communication office | Low | Test alert system weekly. Translate into local languages |
| Resource Allocation | Semi-automated | High - disaster manager decides final allocation | Low - AI math is not transparent | Manual allocation using past experience | Disaster manager and mayor | High | Show AI reasoning to manager. Require written approval. Check if allocation was fair |
| Feedback & Improvement | Manual only | High - humans review mistakes | High - reports are documented | Paper system exists as backup | PAGASA and external auditors | Medium | Create formal error reporting system. Monthly review of mistakes. Fix problems documented |
Main Risks Found
Trustworthiness Strengths
Trustworthiness Gaps
Technical Safeguards (Code & Software)
Process Safeguards (How People Use the System)
Policy Safeguards (Rules & Accountability)
Human-Centered Design Improvements
The Disaster AI system has high potential to save lives. It also has high risk if it fails. Making it trustworthy requires:
The most important principle: Put humans and communities first. Use AI to support human decisions not replace them.
You can now copy and paste these directly into your Google Classroom or Google Docs. Everything uses simple words and short sentences!
I can see your Google Sheets is all set up! You have the Trustworthiness Evaluation Matrix with the right headers. Now you need to fill in the rows with information about your AI system.
Here is the data you can copy and paste into your spreadsheet based on the Disaster Relief AI example I gave you:
Row 2: Data Collection & Sensors
Row 3: Risk Prediction Algorithm
Row 4: Alert Distribution
Row 5: Resource Allocation
Row 6: Feedback & Improvement
Just copy each row into your Google Sheet and you will have a complete Trustworthiness Evaluation Matrix for the Disaster Relief AI system!
can yo just do this
A. Risk Analysis & HCAI Classification
Step 1: Select a documented AI case
Choose a real-world AI system, such as:
• Scholarship ranking system
• Loan approval system
• Social media moderation platform
• Disaster relief targeting system
Step 2: Identify risks
Using Google Docs, list risks related to:
• Reliability: Errors, inconsistent outputs
• Safety: Physical, financial, or psychological harm
• Misuse: Bias, discrimination, or unfair treatment
Step 3: Classify the AI system
Apply the Two-Dimensional HCAI Framework:
• Axis 1: Human vs. AI control (manual, assisted, semi-automated, fully automated)
• Axis 2: Context criticality (low, medium, high)
Step 4: Document potential failure points
Highlight system components that may fail or cause harm, noting severity and likelihood.
also atached is our matrix base it from this
You can send messages while the preview loads