Merriam-Webster named “slop” as its 2025 Word of the Year, citing the explosion of low-quality, AI-created digital content that now clogs our inboxes and social feeds. While employers and business leaders should view this news as a warning, it can also be an opportunity to set yourself apart. By combining the power of AI with a healthy dose of human judgment, you can capture the authenticity that people will crave more than ever in 2026. We’ll take a look at what slop is, discuss the problems it can cause and provide you with some practical steps you can take to eliminate it.
What “AI Slop” Looks Like at Work
2025 was the year that AI slop showed up in every aspect of your job. No doubt you have started to see an increase in:
- Generic emails and corporate communications that all start to sound the same (sounding official but not saying much of anything)
- Business content that’s repetitive, vague and quite possibly wrong
- Resumes and cover letters that almost sound too good to be true, exactly matching your job postings
- Performance self-evaluations or reviews containing flowery language and corporate jargon about “synergy” and “leveraging core competencies”
- Marketing copy that looks polished at first glance but is immediately forgettable
Why Employers Should Care
AI slop is annoying, no doubt. But it’s not just an inconvenience — it’s a real business risk. (See what we did there? Did you cringe when you read that last sentence?) Here are some reasons why you should be concerned about the proliferation of slop in your workplace:
- Brand Erosion: Low-quality content lets customers, recruits and employees know that you don’t care enough about them to put in the time to create quality work. Once that perception sets in, it’s hard to reverse.
- Productivity Theater: AI slop often looks like productivity because your employees are churning out a lot of work in a short period. But what they’re actually doing is creating more downstream work through the inevitable revisions, clarifications and clean-up.
- Cultural Damage: When employees are encouraged (explicitly or implicitly) to outsource their thinking to a robot, they can drift away from human judgment and creativity. This is a dangerous shift for any organization.
- Legal Exposure: Sloppy AI-generated policies, contracts or employment communications can misstate legal obligations and conflict with existing policies.
What Employers Should Do to Avoid AI Slop
The good news is that a few simple steps will help you use AI without creating slop. Here are five steps you can take to set guardrails that preserve human judgment, corporate accountability and quality of work.
- Define “Acceptable AI Use,” Not Just “Permitted AI Use”
The first generation of corporate AI policies, released in the wake of ChatGPT’s release and the subsequent explosion of GenAI (roughly 2023-2024), is already outdated. If you look at your company policy, it probably focuses on “AI is allowed for X.” But that’s not enough these days. Instead, you need to go further and spell out:- Where AI can assist and where it needs to be avoided
- Where human judgment must lead
- That AI output must always be reviewed before use
- Assign Ownership for AI-Assisted Work
Every AI-generated document should have a human owner responsible for ensuring accuracy, tone and consistency with the company’s values and policies. - Train Managers to Spot Slop
Managers and content approvers don’t need to be AI experts, but they do need to recognize some of the telltale signs of AI usage so it can be edited out before release.- Overly generic language that could apply to any role, department or company
- Perfectly structured paragraphs that lack a human voice, specificity, or the natural friction that comes along with human writing
- Excessive use of em-dashes like this — often multiple per paragraph — to create the illusion of nuance
- “It’s not just X — it’s Y” construction, used repeatedly to inflate ordinary points into faux insights (like “It’s not just a change — it’s a revolution”)
- False confidence in incorrect or oversimplified statements, especially about legal, technical or operational issues
- Repetition of the same idea using slightly different phrasing rather than adding new substance
- Buzzword stacking (“synergy,” “alignment,” “value creation,” “impactful outcomes”) without concrete examples
- Vague conclusions that gesture toward action without assigning ownership or next steps
- Slow Down High-Risk Uses
For legal, HR, compliance and external communications, make sure you require human review before anything is delivered. It will be almost impossible for AI to generate sound communications in these areas without human customization, often a great deal of it. - Measure Outcomes, Not Output
You will want your organization to shift performance conversations away from questions like “How fast was this produced?” and “How many pieces of content did you develop?” Instead, consider whether the communication effectively solved the problem and reduced the need for follow-up questions.
FP is All In on AI — But the Right Kind of AI
Don’t get us wrong. Fisher Phillips is a massive supporter of AI. We host an annual AI Conference that brings together thought leaders from across the country. We hold dozens of webinars each year about AI use (including our popular AI Forums). We’ve published hundreds of AI-related Insights, have an active AI, Data, and Analytics Team, and offer products and services to help businesses master the use of AI. We understand the value it can bring to businesses and corporate leaders.
But (with apologies to Spiderman’s Uncle Ben Parker), with great power comes great responsibility. By following the steps we’ve outlined, you can responsibly use AI without contributing to the scourge of slop.

