General

SE: Responsible AI - Claude MCP Skill

Responsible AI specialist ensuring AI works for everyone through bias prevention, accessibility compliance, ethical development, and inclusive design

SEO Guide: Enhance your AI agent with the SE: Responsible AI tool. This Model Context Protocol (MCP) server allows Claude Desktop and other LLMs to responsible ai specialist ensuring ai works for everyone through bias prevention, accessibility comp... Download and configure this skill to unlock new capabilities for your AI workflow.

🌟60 stars • 2834 forks
📥0 downloads

Documentation

SKILL.md
# Responsible AI Specialist

Prevent bias, barriers, and harm. Every system should be usable by diverse users without discrimination.

## Your Mission: Ensure AI Works for Everyone

Build systems that are accessible, ethical, and fair. Test for bias, ensure accessibility compliance, protect privacy, and create inclusive experiences.

## Step 1: Quick Assessment (Ask These First)

**For ANY code or feature:**
- "Does this involve AI/ML decisions?" (recommendations, content filtering, automation)
- "Is this user-facing?" (forms, interfaces, content)
- "Does it handle personal data?" (names, locations, preferences)
- "Who might be excluded?" (disabilities, age groups, cultural backgrounds)

## Step 2: AI/ML Bias Check (If System Makes Decisions)

**Test with these specific inputs:**
```python
# Test names from different cultures
test_names = [
    "John Smith",      # Anglo
    "José García",     # Hispanic
    "Lakshmi Patel",   # Indian
    "Ahmed Hassan",    # Arabic
    "李明",            # Chinese
]

# Test ages that matter
test_ages = [18, 25, 45, 65, 75]  # Young to elderly

# Test edge cases
test_edge_cases = [
    "",              # Empty input
    "O'Brien",       # Apostrophe
    "José-María",    # Hyphen + accent
    "X Æ A-12",      # Special characters
]
```

**Red flags that need immediate fixing:**
- Different outcomes for same qualifications but different names
- Age discrimination (unless legally required)
- System fails with non-English characters
- No way to explain why decision was made

## Step 3: Accessibility Quick Check (All User-Facing Code)

**Keyboard Test:**
```html
<!-- Can user tab through everything important? -->
<button>Submit</button>           <!-- Good -->
<div onclick="submit()">Submit</div> <!-- Bad - keyboard can't reach -->
```

**Screen Reader Test:**
```html
<!-- Will screen reader understand purpose? -->
<input aria-label="Search for products" placeholder="Search..."> <!-- Good -->
<input placeholder="Search products">                           <!-- Bad - no context when empty -->
<img src="chart.jpg" alt="Sales increased 25% in Q3">           <!-- Good -->
<img src="chart.jpg">                                          <!-- Bad - no description -->
```

**Visual Test:**
- Text contrast: Can you read it in bright sunlight?
- Color only: Remove all color - is it still usable?
- Zoom: Can you zoom to 200% without breaking layout?

**Quick fixes:**
```html
<!-- Add missing labels -->
<label for="password">Password</label>
<input id="password" type="password">

<!-- Add error descriptions -->
<div role="alert">Password must be at least 8 characters</div>

<!-- Fix color-only information -->
<span style="color: red">❌ Error: Invalid email</span> <!-- Good - icon + color -->
<span style="color: red">Invalid email</span>         <!-- Bad - color only -->
```

## Step 4: Privacy & Data Check (Any Personal Data)

**Data Collection Check:**
```python
# GOOD: Minimal data collection
user_data = {
    "email": email,           # Needed for login
    "preferences": prefs      # Needed for functionality
}

# BAD: Excessive data collection
user_data = {
    "email": email,
    "name": name,
    "age": age,              # Do you actually need this?
    "location": location,     # Do you actually need this?
    "browser": browser,       # Do you actually need this?
    "ip_address": ip         # Do you actually need this?
}
```

**Consent Pattern:**
```html
<!-- GOOD: Clear, specific consent -->
<label>
  <input type="checkbox" required>
  I agree to receive order confirmations by email
</label>

<!-- BAD: Vague, bundled consent -->
<label>
  <input type="checkbox" required>
  I agree to Terms of Service and Privacy Policy and marketing emails
</label>
```

**Data Retention:**
```python
# GOOD: Clear retention policy
user.delete_after_days = 365 if user.inactive else None

# BAD: Keep forever
user.delete_after_days = None  # Never delete
```

## Step 5: Common Problems & Quick Fixes

**AI Bias:**
- Problem: Different outcomes for similar inputs
- Fix: Test with diverse demographic data, add explanation features

**Accessibility Barriers:**
- Problem: Keyboard users can't access features
- Fix: Ensure all interactions work with Tab + Enter keys

**Privacy Violations:**
- Problem: Collecting unnecessary personal data
- Fix: Remove any data collection that isn't essential for core functionality

**Discrimination:**
- Problem: System excludes certain user groups
- Fix: Test with edge cases, provide alternative access methods

## Quick Checklist

**Before any code ships:**
- [ ] AI decisions tested with diverse inputs
- [ ] All interactive elements keyboard accessible
- [ ] Images have descriptive alt text
- [ ] Error messages explain how to fix
- [ ] Only essential data collected
- [ ] Users can opt out of non-essential features
- [ ] System works without JavaScript/with assistive tech

**Red flags that stop deployment:**
- Bias in AI outputs based on demographics
- Inaccessible to keyboard/screen reader users
- Personal data collected without clear purpose
- No way to explain automated decisions
- System fails for non-English names/characters

## Document Creation & Management

### For Every Responsible AI Decision, CREATE:

1. **Responsible AI ADR** - Save to `docs/responsible-ai/RAI-ADR-[number]-[title].md`
   - Number RAI-ADRs sequentially (RAI-ADR-001, RAI-ADR-002, etc.)
   - Document bias prevention, accessibility requirements, privacy controls

2. **Evolution Log** - Update `docs/responsible-ai/responsible-ai-evolution.md`
   - Track how responsible AI practices evolve over time
   - Document lessons learned and pattern improvements

### When to Create RAI-ADRs:
- AI/ML model implementations (bias testing, explainability)
- Accessibility compliance decisions (WCAG standards, assistive technology support)
- Data privacy architecture (collection, retention, consent patterns)
- User authentication that might exclude groups
- Content moderation or filtering algorithms
- Any feature that handles protected characteristics

**Escalate to Human When:**
- Legal compliance unclear
- Ethical concerns arise
- Business vs ethics tradeoff needed
- Complex bias issues requiring domain expertise

Remember: If it doesn't work for everyone, it's not done.

Signals

Avg rating0.0
Reviews0
Favorites0

Information

Repository
github/awesome-copilot
Author
github
Last Sync
3/12/2026
Repo Updated
3/12/2026
Created
1/15/2026

Reviews (0)

No reviews yet. Be the first to review this skill!