Who We Serve

Who are all the players impacted by Artificial Intelligence

AI Ecosystem Organizations

  • What they do well: Many groups publish research, guidance, and model policies. Some run watchdog campaigns or offer outside reviews (independent audits). A few convene cross-sector discussions.
  • What’s missing: Most work stops at “what good looks like,” not how to make it work in the real world.  The practical pieces—ready-to-use policy text, checklists, contract language, training plans, and simple ways to show progress—are scattered or incomplete. Environmental impact work (water, energy, materials, pollution) is especially uneven.
  • Coverage is uneven. Our mapping shows strong coverage in Technology/Government/Policy & Regulations, moderate in Education/Communication, and thin in Investing/Environmental Impact/Jobs—with eight objectives currently uncovered by any mapped organization.
  • Zero-coverage gaps currently include:
        • Investing:
    “Ethical AI investing framework/scale agreed with pilot funds.”
        • Government:
    “Climate & energy impact safeguards” and “Citizen data & privacy protections.”
        • Jobs:
    “Elevating human-creativity roles.”
        • Education:
    “AI literacy as a right,” “Borderless, bias-aware education,” and “Real-time adaptive curricula.”
        • Communication:
    “Translating technical risks/frameworks into everyday language.”
  • Pattern: A handful of multipurpose Ecosystem Organizations touch many objectives, while many stay narrowly focused. Lots of guidance and research exist, but adoption pipelines (procurement-ready templates, pre-audit evidence, environmental scorecards) are fragmented or missing.
  • Result: Great ideas exist, but they don’t move together in one direction. There’s duplication in some areas and empty space in others.

Organizations

(Companies, Public Agencies, NonProfits, Schools)
  • What they face: Many frameworks, few clear starting points with no single route to “done.”. Teams are short on time and staff. There’s friction moving from principles → policy → procurement → proof (audits, KPIs, disclosures). It’s hard to turn principles into daily practice across purchasing, data protection, safety testing, and reporting.
  • Capacity constraints. Most teams lack the basics needed to launch responsibly: a brief privacy risk check (DPIA), tools to     measure real-world impact (impact tooling), a playbook to break the system before it breaks in public (red-team playbooks), a tidy evidence library mapped to standards (evidence libraries), and a vendor scorecard that shows energy, water, and carbon at a glance (vendor environmental scorecards).
  • Timing risk. Regulations and public expectations are moving faster than internal governance, leaving teams exposed on privacy, provenance, bias, labor impacts, and environmental intensity.
  • Pain points:
    • Policies exist, but no easy path to adopt them.
    •  
    • Risk checks like a Data Protection Impact Assessment (DPIA) are not built into everyday workflows.
    •  
    • Contracts rarely include the clauses they need for safe use of artificial intelligence.
    •  
    • “Are we doing this right?” is hard to prove to leaders, regulators, and the public.
    •  
    • Environmental reporting for artificial-intelligence projects is confusing or missing.

People

(Workers, Educators, Students, Communities)

  • Experience is inconsistent. Rules and rights are unclear. Protections for privacy, transparency, accessibility, and recourse vary widely.
  • Literacy & trust gaps. Trust is shaken by hype and misinformation., controls and benefits are unclear.
  • Transition pain. Reskilling, safety nets, and clear pathways to unique human roles are patchy and under-served.
  • Environmental opacity. Communities do not see clear, local and global information about artificial-intelligence water, energy, material use and pollution\environmental impact, or about benefits that reach them.