California Education Department Launches AI Workgroup
The California Department of
Education (CDE) has officially launched its first statewide Artificial
Intelligence in Education Workgroup. This group will focus on how schools
can use AI responsibly in classrooms while protecting student privacy and
keeping teachers at the center of decision-making.
Led by State Superintendent of
Public Instruction Tony Thurmond, the initiative is part of
California’s broader effort to set clear guidelines on AI use in schools. The
move follows the passage of Senate Bill 1288 (SB 1288) in
2024, which requires the state to take a careful, structured approach to
artificial intelligence in education.
In short, California wants to
make sure AI supports learning rather than replacing teachers, and that its use
is safe, ethical, and equitable.
Why
a Workgroup on AI?
AI technologies are spreading
rapidly in education, from automated grading tools and tutoring
apps to student data platforms. While these tools offer
exciting possibilities, they also raise concerns.
- Will AI increase student engagement or widen the gap between those
with access and those without?
- How can schools protect student data when using AI tools?
- How do educators maintain academic integrity when tools can
generate essays and homework answers within seconds?
In a recent press statement,
Thurmond emphasized the need for clear guidelines on AI in schools, noting that
technology should strengthen—rather than replace—the essential role of
educators.”
Who
Is Involved?
The workgroup gathered for its
first meeting on August 29, 2025, at the CDE headquarters in Sacramento,
bringing together a diverse group of voices.
- Teachers (at least half of the group)
- Students
- School administrators
- Classified staff (such
as support staff and aides)
- Higher education representatives
- Industry experts
The wide mix of members
highlights an important point: decisions about using AI in schools should not
be left only to policymakers or tech companies. Instead, teachers, students,
and school staff who experience classrooms every day need to shape how AI is
introduced.
As Thurmond explained, “We
want to make sure the people most affected by these policies have a real say in
creating them.”
What
Does Senate Bill 1288 Require?
1 . Senate
Bill 1288, written by Senator Josh Becker and co-sponsored by Thurmond, sets out the key
responsibilities for the AI in Education Workgroup. These include:
o
Reviewing
how AI is used in schools now and what the future might look like – This means identifying which tools are
already in use, their costs, who owns them, and how they are affecting both
teaching and learning.
o
Which AI tools are already being
used?
o
What are their costs, ownership
models, and long-term impacts?
o
How are they helping—or
harming—teachers and students?
2.
Host public
meetings
o
At least six meetings must be
held.
o
Feedback will be gathered from
students, parents, teachers, and community members.
3.
Develop
clear guidance for school districts
o
Policies on academic
integrity (preventing plagiarism and misuse).
o
Rules around student and
teacher privacy.
o
Provide examples of how AI
should and should not be used in schools
4.
Develop a
model policy for districts and charter schools
o
A statewide guideline that
schools can follow or adjust to fit their needs.
o
It will also address equity,
making sure every student has fair access to AI tools and resources.Create a
model policy for districts and charter schools
Timeline
for Implementation
The legislation outlines a clear
schedule:
- By January 1, 2026, the group will issue its first set of guidance.
- July 1, 2026 –
A full model policy on AI use must be published.
- 2027 – A final report goes to the state
Legislature.
- 2031 – The statute establishing the
workgroup will officially dissolve.
This step-by-step schedule gives
schools time to prepare, while ensuring the conversation moves forward without
unnecessary delays.
Why
This Matters for Students and Educators
AI can feel overwhelming. Some
teachers worry it may undermine traditional teaching, while others see it as a
powerful partner. Students are also divided—some use AI for brainstorming or
practice, while others are tempted to rely on it too heavily for assignments.
The creation of this workgroup
is significant because it:
- Puts teachers’ voices first –
ensuring technology doesn’t replace but supports them.
- Sets consistent statewide standards – so every district, from large cities to small towns, has
clear guidance.
- Balances innovation with responsibility – encouraging new learning opportunities without ignoring
risks like data misuse.
By starting this conversation
now, California hopes to prepare schools for both the opportunities and
challenges AI will bring over the next decade.
Looking
Ahead
The launch of the AI in
Education Workgroup makes California one of the first states in the U.S. to
take a legislatively mandated statewide approach to artificial
intelligence in classrooms.
If successful, this initiative
could serve as a model for other states. With AI becoming an
unavoidable part of daily life, giving educators, parents, and students a seat
at the table is not just smart policy—it’s essential for building trust.
As the first set of
recommendations is expected in early 2026, schools and districts across
California will be watching closely. The outcomes could shape how millions of
students interact with technology for years to come.
For
a deeper look at how AI is transforming classrooms globally, see this UNESCO report on AI in education.
You
can also read more about education technology trends on our
site’s EdTech resources page.
Conclusion
The California Department of
Education’s AI Workgroup is not just about technology—it’s about people. It
recognizes that AI can be a helpful tool, but only if used thoughtfully,
ethically, and in ways that keep teachers and students at the center. By creating
statewide standards, California is taking a careful step forward, ensuring that
innovation in classrooms remains balanced with equity, privacy, and human
connection.
