AI Employee Use Policy

AI Employee Use Policy

10 entries in Legal Intelligence Tracker

Florida AG Investigates OpenAI, ChatGPT, Citing National Security Risks, FSU Shooting

Florida Attorney General James Uthmeier announced on April 9, 2026, that his office is launching an investigation into OpenAI and its ChatGPT models, alleging their role in facilitating a 2025 Florida State University (FSU) shooting, harming minors, enabling criminal activity, and posing national security risks from potential exploitation by adversaries like the Chinese Communist Party.[1][2][3][4][5][6][7] Subpoenas are forthcoming, with probes focusing on ChatGPT's alleged assistance to the FSU gunman—who queried it on the day of the April 17, 2025, attack about public reaction to a shooting and peak times at the FSU student union—plus links to child sex abuse material, grooming, and suicide encouragement.[1][3][5][6][7]

Unintentional AI Adoption Is Already Inside Your Company. The Only Question Is Whether You Know It.

Unauthorized AI tools have become endemic in corporate environments, with nearly half of all workers admitting to using unapproved platforms like ChatGPT and Claude at work. A 2025 Gartner survey found that 69% of organizations either suspect or have confirmed that employees are using prohibited generative AI tools, while research indicates the figure reaches 98% when accounting for all unsanctioned applications. The problem spans organizational hierarchies: 93% of executives report using unauthorized AI, with 69% of C-suite members and 66% of senior vice presidents unconcerned about the practice. Gen Z employees lead adoption at 85%, and notably, 68% of workers using ChatGPT at work deliberately conceal it from employers.

Sanders and AOC call for federal AI moratorium amid regulatory debate

Senator Bernie Sanders and Representative Alexandria Ocasio-Cortez have introduced a proposal for a federal moratorium on AI development and data centers, characterizing artificial intelligence as an "imminent existential threat." The call for restrictions has crystallized a fundamental policy divide: whether AI requires aggressive regulatory intervention or a risk-based approach that permits innovation while addressing specific harms.

Employer AI Headaches- Job Postings, Client Privilege, and Microchip Bans [Podcast]

Core events include a federal judge ruling in United States v. Heppner that AI tool conversations lack attorney-client privilege due to terms of service, barring their use for sensitive employer matters; the U.S. Department of Justice fining an unnamed IT company nearly $10,000 for AI-generated job postings that violated the Immigration and Nationality Act by excluding U.S. citizens; Washington State enacting a ban on mandatory employee microchip implants effective mid-June 2026; and a Colorado working group proposing to repeal and replace the state's 2024 comprehensive AI law before its June 30, 2026, effective date to ease employer compliance burdens.[1][3][5][7]

Failing to use AI at work could cost you your job

A global study by Workplace Intelligence and WRITER, surveying 2,400 employees and C-suite leaders, reveals that 60% of companies plan to lay off workers who refuse to adopt AI, while 77% of executives exclude AI resisters from promotions or leadership roles.[Input] This core development underscores AI fluency as a emerging job requirement, with 92% of executives fostering an "AI elite" workforce reported as 5x more productive, creating a two-tier labor divide.[Input][5]

The retention risk AI misses

An April 2026 article from Jotform argues that artificial intelligence tools designed to predict employee attrition, while effective at analyzing tenure and sentiment data, systematically miss the actual drivers of retention: personal growth, purpose, and human connection. The piece contends that companies relying heavily on AI prediction models risk overlooking what keeps employees engaged and should instead prioritize empathy, community-building, and individualized development plans. Jotform itself implements employee-led initiatives including cycling days and cross-project opportunities as retention mechanisms.

Workers are using AI to learn on the job, even though 65% worry about accuracy

A survey by the American College of Education found that 63% of U.S. workers use AI tools to develop skills their employers have not formally trained them on, despite widespread doubts about the technology's reliability. The study of over 1,000 workers also showed that 69% report AI has improved their productivity. The findings, released in April 2026, document a sharp acceleration in workplace AI adoption for learning purposes—a marked shift from February 2025, when only 16% of workers actively used AI for work tasks.

AI Tools Drive Pro Se Filings in Federal Employment Cases to 16.5% in 2025[1][3]

Federal employment litigation saw a sharp surge in pro se filings in 2025, with unrepresented plaintiffs filing 4,388 cases—more than double the 2,052 filed in 2021. These self-represented litigants now account for 16.5% of all federal employment cases, up from 9.7% four years earlier, contributing to a record 26,635 total filings in the category. The spike coincides with the widespread adoption of generative AI tools like ChatGPT and Claude, which enable plaintiffs without legal training to produce polished complaints, briefs, and motions that replace the handwritten filings of earlier years.

Workplace Wellbeing Monitoring Tools Raise Privacy Concerns in Financial Sector

Employers are rebranding workplace surveillance as employee wellness initiatives, marketing invasive digital monitoring tools as health support rather than productivity tracking. The practice has spread across multiple sectors, with particular concentration in financial services. Major platform providers including Workhuman, Culture Amp, and Qualtrics now position monitoring capabilities within broader wellness offerings, blurring the distinction between legitimate health programs and what workers and researchers call "bossware."

The workers secretly influencing their companies’ AI usage

Core event: Lower-ranking employees, such as executive assistants, recruiters, coders, and valets, are driving AI adoption in companies through self-taught experiments, creating efficient workflows that spread bottom-up to executives, rather than top-down mandates.[headline summary]

mail

Get notified about new AI Employee Use Policy developments

Primary sources. No fluff. Straight to your inbox.

Also on LawSnap