Policy & Leadership · Blueprints

School AI Policy Examples That Actually Work

By Shawn Pecore April 2, 2026 7 min read
School AI policy examples that work in 2026: Gwinnett County, Chicago, Boston, and Peninsula School District frameworks

While most schools are either banning AI or waiting to see what happens, a smaller group of districts built functioning policy frameworks and are running them in live classrooms. Here is what they built and what they have in common.

  • All four districts replaced prohibition with structured integration, treating AI literacy as a curriculum goal rather than an institutional threat.
  • All four age-gate access in some form, recognising that the appropriate policy for an 8-year-old is not the same as for a 17-year-old.
  • All four shift the assessment focus from the polished final product to the documented process of thinking.
  • The common thread: teacher agency over administrative surveillance, with clear guidelines rather than blanket rules.

What the Working Policies Share

The districts with functional AI policies did not converge on identical solutions. Their frameworks differ by structure, emphasis, and institutional context. But across all of them, the same underlying decisions produced the outcomes that distinguish them from districts still cycling through bans and detection software.

First: they treated AI as a literacy problem rather than a compliance problem. The question they started with was not "how do we stop students from using AI?" but "what do we need students to know and be able to do with AI?" That reframe changes everything downstream, from assessment design to professional development to the specific language in the policy document.

Second: they involved teachers in the drafting process. None of these policies was handed down from administration as a finished document. In each case, classroom educators shaped the substance of what was written. That involvement means the policies reflect what is actually possible in a classroom rather than what sounds reasonable in an administrative meeting.

Third: they accepted that assessment has to change. A policy that addresses AI without addressing how student knowledge is evaluated has not addressed the actual problem. Each of these districts redesigned at least some of their assessment toward process documentation, oral defence, or assignment structures that require content the AI cannot fabricate.

Gwinnett County Public Schools, Georgia: The K-12 Progression

Gwinnett County built the most structurally ambitious AI policy of any large US district, and they started with the most useful question: what should a student know about AI at each stage of their development, and how does that build year on year?

The answer is a learning progression that begins in Grade 1. First graders are not using AI tools. They are learning the foundational concept of sequential logic by describing the specific steps required to assemble a sandwich: a task that teaches procedural thinking without any technology. By middle school, students are using Python to program charts that decipher genetic codes. At Seckinger High School, a rigorous three-course AI pathway is available for students interested in building AI tools, not just using them. Edutopia, 2026

The district policy explicitly permits AI for research, foreign language development, and brainstorming, with the AI Assessment Scale framework built into how specific assignments are labelled. Students know before they start an assignment what AI use is permitted on it. The ambiguity that drives most academic misconduct is removed at the assignment level, not managed through retrospective detection.

The Gwinnett model works because it treats AI as a subject to be taught progressively, not a tool to be permitted or denied uniformly across twelve grade levels with vastly different developmental contexts.

Chicago Public Schools: Age-Gated Access

Chicago Public Schools released its AI Guidebook in July 2024, making it one of the earliest large district policy frameworks in the US. The distinguishing feature of the Chicago approach is age-gating: the permitted level of AI autonomy scales explicitly with developmental stage rather than applying a uniform standard across 300,000 students. Playlab, 2026

Elementary school students use AI only under direct teacher supervision, for specific structured tasks like character interviews where the teacher controls the interaction and debriefs the content. Middle school students access AI for virtual experiments and collaborative tasks, with teacher oversight maintained. High school students have independent access for essay feedback, synthesis of research, and exploratory work.

The logic is developmental: an 8-year-old does not yet have the critical evaluation skills to assess AI output independently. A 17-year-old can develop those skills and should be doing so in a supported classroom environment rather than encountering AI for the first time in an unsupported workplace. Chicago's framework builds the scaffolding that makes independent high school use responsible rather than risky.

The age-gating approach prevents early-childhood cognitive debt without banning AI from the educational environment entirely. Young children build foundational skills in structured, supervised AI interactions rather than developing an unexamined reliance on the tool before they have the critical thinking skills to question its output.

Boston Public Schools: The Five-Pillar Framework

Boston Public Schools released its AI guidance in August 2025, and the framework reflects a deliberate philosophy: AI policy in a school district is not primarily a technology decision. It is an educational values decision.

The five pillars are Academic Excellence, Cybersecurity, Ethics, Bias Acknowledgment, and Human Oversight. Playlab, 2026 Each one reflects a specific commitment. Academic Excellence means AI use must serve genuine learning outcomes, not just grade outcomes. Cybersecurity means students and teachers are protected through vendor vetting and data privacy compliance. Ethics and Bias Acknowledgment mean AI tools are examined for what they encode and what populations they disadvantage. Human Oversight means a person is always accountable for what AI produces.

Critically, the Boston policy treats AI use as a case-by-case decision made by teachers based on specific learning goals rather than a blanket mandate applied district-wide. This is the teacher-agency principle in practice. The policy provides a framework of values and a list of approved, vetted tools. The judgment about whether AI serves the learning goal of a specific assignment on a specific day with a specific class is left to the professional in the room.

The district maintains an approved tools list with a strict vendor vetting process. No tool enters the classroom without being evaluated for data privacy compliance. This protects students and removes the ambiguity teachers face when well-intentioned colleagues use unapproved tools with student data.

Peninsula School District, Washington: Human-AI-Human

Peninsula School District's framework is the most conceptually clear of the four. Every AI interaction in a school context must begin with human inquiry and close with human reflection. The AI occupies the middle of the process. It does not start it and it does not conclude it. Washington OSPI, 2024

The practical application is straightforward. A student cannot open an AI tool and ask it to generate an essay plan. They must first do the human work of forming a question, identifying what they already know and do not know, and articulating what they need. The AI then assists with specific, bounded tasks. The student then does the human work of critically evaluating the AI output, integrating it with their own thinking, and taking responsibility for what they submit.

The framework integrates Universal Design for Learning principles, which means it is designed from the start to serve students with varying abilities and learning needs. AI tools are positioned as accessibility supports and cognitive scaffolds, not shortcuts. The assessment focus is on the process documentation: what the student did before the AI, what they asked the AI, what they changed or rejected from the AI output, and what they concluded independently.

The Human-AI-Human model sidesteps the detection problem entirely. There is nothing to detect when the policy requires evidence of the human-authored components at every stage. A student who submits without the process documentation has failed to meet the assignment requirements regardless of whether a detector flags the content.

What to Take from These Examples

No district policy can be copied wholesale into a different institutional context. Gwinnett's K-12 progression works partly because it was built with the specific community and staffing of Gwinnett County in mind. But the structural decisions these districts made are transferable.

Age-gate access rather than applying uniform rules across K-12. Build AI literacy into curriculum progressively rather than treating it as a one-time training event. Give teachers clear, specific guidance at the assignment level rather than vague principles. Evaluate AI tools for bias and data privacy before deploying them. Design assessment to evaluate process rather than polished final product. Keep a human accountable at the start and end of every AI-assisted task.

None of those decisions require a large budget or a dedicated technology department. They require an institutional willingness to accept that the old approach is not working and to involve the people closest to the problem in building the replacement.

For the full analysis of what most schools are doing instead and why it is failing, see What Schools Are Getting Wrong About AI Policy. For a look at who gets hurt most by the failure, see The AI Policy Equity Gap: Who Gets Left Behind.

FAQ

Gwinnett County, Georgia has a K-12 AI learning progression from foundational logic in Grade 1 to a three-course high school pathway. Chicago Public Schools uses age-gated access by developmental stage. Boston Public Schools operates on a five-pillar framework with teacher-led case-by-case decisions and a vetted tools list. Peninsula School District in Washington uses the Human-AI-Human model requiring human inquiry and human reflection to frame every AI interaction.

All four districts share: AI literacy treated as a curriculum goal rather than a compliance problem, teacher agency over administrative surveillance, age-appropriate access rather than uniform rules, assessment focused on process documentation rather than polished final product, and a vendor vetting procedure for data privacy.

Peninsula School District's framework requires every AI interaction to begin with human inquiry and close with human reflection. The student forms a question and articulates what they know before opening any AI tool. After AI assistance, the student critically evaluates the output, integrates it with their own thinking, and takes responsibility for what they submit. The process documentation is part of the assessment.

Elementary students access AI only under direct teacher supervision for specific structured tasks. Middle school students use AI for experiments and collaborative work with teacher oversight. High school students have independent access for essay feedback, research synthesis, and exploratory tasks. The permitted autonomy scales with the student's developmental stage and critical evaluation capacity.

Yes. The structural decisions these districts made are transferable without requiring large budgets or technology departments: age-gate access, build AI literacy into curriculum progressively, give teachers clear assignment-level guidance, evaluate tools for bias and data privacy, design assessments for process rather than product, and keep a human accountable at the start and end of every AI-assisted task.

Sources

  1. Edutopia. How Forward-Thinking Schools Are Shaping the Future of AI in Education. 2026. edutopia.org
  2. Playlab. District and School AI Policies. 2026. learn.playlab.ai
  3. Washington OSPI. A Practical Guide for Implementing AI in the Classroom. 2024. ospi.k12.wa.us
  4. Education Commission of the States. Artificial Intelligence Technologies in Education. July 2025. ecs.org
  5. Education Week. 5 Best Practices for Crafting a School AI Policy. October 2025. edweek.org
About the Author

Shawn Pecore is an educator, scientist, and author with classroom and global consulting experience. He researches, writes, and discusses current issues in AI in education facing educators, parents, and students. Follow along on Substack at @schoollyai for new posts and updates.

Shawn also writes about where education is heading and publishes children's science books through the MEYE Science Series. Visit shawnpecore.com and follow him on Substack at @shawnpecore.