https://www.instagram.com/reel/DO9MwTHCoR_/?igsh=MTZybml2NDB...
The screenshots/videos of them doing it are pretty wild, and insane they are editing creators' uploads without consent!
To me, this clearly looks like a case of a very high compression ratio with the motion blocks swimming around on screen. They might have some detail enhancement in the loop to try to overcome the blockiness which, in this case, results in the swimming effect.
It's strange to see these claims being taken at face value on a technical forum. It should be a dead giveaway that this is a compression issue because the entire video is obviously highly compressed and lacking detail.
Such as?
This seems like such an easy thing for someone to document with screenshots and tests against the content they uploaded.
So why is the top voted comment an Instagram reel of a non-technical person trying to interpret what's happening? If this is common, please share some examples (that aren't in Instagram reel format from non-technical influencers)
It's difficult for me to read this as anything other than dismissing this person's views as being unworthy of discussing because they are are "non-technical," a characterization you objected to, but if you feel this shouldn't be the top level comment I'd suggest you submit a better one.
Here's a more detailed breakdown I found after about 15m of searching, I imagine there are better sources out there if you or anyone else cares to look harder: https://www.reddit.com/r/youtube/comments/1lllnse/youtube_sh...
To me it's fairly subtle but there's a waxy texture to the second screenshot. This video presents some more examples, some of them have are more textured: https://www.youtube.com/watch?v=86nhP8tvbLY
Perhaps a useful analogy is "breaking userspace." It's important to correctly diagnose a bug breaking userspace to ship a fix. But it's a bug if its a change that breaks userspace workflows, full stop. Whether it met the letter of some specification and is "correct" in that sense doesn't matter.
If you change someone's appearance in your post processing to the point it looks like they've applied a filter, your post processing is functionally a filter. Whether you intended it that way doesn't change that.
If your compression pipeline gives people anime eyes because it's doing "detail enhancement", your compression pipeline is also a filter. If you apply some transformation to a creator's content, and then their viewers perceive that as them disingenuously using a filter, and your response to their complaints is to "well actually" them about whether it is a filter or a compression artifact, you've lost the plot.
To be honest, calling someone "non-technical" and then "well actually"ing them about hair splitting details when the outcome is the same is patronizing, and I really wish we wouldn't treat "normies" that way. Regardless of whether they are technical, they are living in a world increasingly intermediated by technology, and we should be listening to their feedback on it. They have to live with the consequences of our design decisions. If we believe them to be non-technical, we should extend a lot of generosity to them in their use of terminology, and address what they mean instead of nitpicking.
I'm not critiquing their opinion that the result is bad. I also said the result was bad! I was critiquing the fact that someone on HN was presenting their non-technical analysis as a conclusive technical fact.
Non-technical is describing their background. It's not an insult.
I will be the first to admit I have no experience or knowledge in their domain, and I'm not going to try to interpret anything I see in their world.
It's a simple fact. This person is not qualified to be explaining what's happening, yet their analysis was being repeated as conclusive fact here on a technical forum
I don't really see where you said the output was "bad," you said it was a compression artifact which had a "swimming effect", but I don't really see any acknowledgement that the influencer had a point or that the transformation was functionally a filter because it changed their appearance above and beyond losing detail (made their eyes bigger in a way an "anime eyes" filter might).
If I've misread you I apologize but I don't really see where it is I misread you.
"Non-technical" isn't an insult.
What you call "well actually"ing is well within limits on a technical forum.
I agreed that the output was bad! I'm not dismissing their concerns, I was explaining that their analysis was not a good technical explanation for what was happening.
This is going to be a huge legal fight as the terms of service you agree to on their platform is “they get to do whatever they want” (IANAL). Watch them try to spin this as “user preference” that just opted everyone into.
"Meta has been doing this; when they auto-translate the audio of a video they are also adding an Al filter to make the mouth of who is speaking match the audio more closely. But doing this can also add a weird filter over all the face."
I don't know why you have to get into conspiracy theories about them applying different filters based on the video content, that would be such a weird micro optimization why would they bother with that
Excessive smoothing can be explained by compression, sure, but that's not the issue being raised there.
Video compression operates on macroblocks and calculates motion vectors of those macroblocks between frames.
When you push it to the limit, the macroblocks can appear like they're swimming around on screen.
Some decoders attempt to smooth out the boundaries between macroblocks and restore sharpness.
The giveaway is that the entire video is extremely low quality. The compression ratio is extreme.
Neural compression wouldn't be like HVEC, operating on frames and pixels. Rather, these techniques can encode entire features and optical flow, which can explain the larger discrepancies. Larger fingers, slightly misplaced items, etc.
Neural compression techniques reshape the image itself.
If you've ever input an image into `gpt-image-1` and asked it to output it again, you'll notice that it's 95% similar, but entire features might move around or average out with the concept of what those items are.
I don't think that's actually what's up, but I don't think it's completely ruled out either.
It looks like they're compressing the data before it gets further processed with the traditional suite of video codecs. They're relying on the traditional codecs to serve, but running some internal first pass to further compress the data they have to store.
Seriously?
Then why is nobody in this thread suggesting what they're actually doing?
Everyone is accusing YouTube of "AI"ing the content with "AI".
What does that even mean?
Look at these people making these (at face value - hilarious, almost "cool aid" levels of conspiratorial) accusations. All because "AI" is "evil" and "big corp" is "evil".
Use occam's razor. Videos are expensive to store. Google gets 20 million videos a day.
I'm frankly shocked Google hasn't started deleting old garbage. They probably should start culling YouTube of cruft nobody watches.
To solve this problem of adding compute heavy processing to serving videos, they will need to cache the output of the AI, which uses up the storage you say they are saving.
Google has already matched H.266. And this was over a year ago.
They've probably developed some really good models for this and are silently testing how people perceive them.
Though there is a LOT of room to subtly train many kinds of lossy compression systems, which COULD still imply they're doing this intentionally. And it looks like shit.
A legal experiment for sure. Hope everyone involved can clear their schedules for hearings in multiple jurisdictions for a few years.
That doesn't include all of the transcoding and alternate formats stored, either.
People signing up to YouTube agree to Google's ToS.
Google doesn't even say they'll keep your videos. They reserve the right to delete them, transcode them, degrade them, use them in AI training, etc.
It's a free service.
It is bad enough we can deepfake anyone. If we also pretend it was uploaded by you the sky is the limit.
And what for? Tiktok creators already generate content for them
And over time the AI content will improve enough where it becomes impossible and then the Great AI Swappening will occur.
There’s already popular subreddits (something blursed ai I think) where people upload this type of content and it’s getting decent engagement it seems
ffmpeg -i source.mkv -i suspect.mkv -filter_complex "blend=all_mode=difference" diff_output.mkv
I saw these claims before but still have not found someone to show a diff or post the source for comparison. It would be interesting.
edit: here's the effect I'm talking about with lossy compression and adaptive quantization: https://cloudinary.com/blog/what_to_focus_on_in_image_compre...
The result is smoothing of skin, and applied heavily on video (as Youtube does, just look for any old video that was HD years ago) would look this way
That would presumably be an easy smoking gun for some content creator to produce.
There are heavy alterations in that link, but having not seen the original, and in this format it's not clear to me how they compare.
These people are having a moral crusade against an unannounced Google data compression test thinking Google is using AI to "enhance their videos". (Did they ever stop to ask themselves why or to what end?)
This level of AI paranoia is getting annoying. This is clearly just Google trying to save money. Not undermine reality or whatever vague Orwellian thing they're being accused of.
https://blog.metaphysic.ai/what-is-neural-compression/
Instead of artifacts in pixels, you'll see artifacts in larger features.
https://arxiv.org/abs/2412.11379
Look at figure 5 and beyond.
> This level of AI paranoia is getting annoying.
Lets be straight here, AI paranoia is near the top of the most propagated subjects across all media right now, probably for worse. If it's not "Will you ever have a job again!?" it's "Will your grandparents be robbed of their net worth!?" or even just "When will the bubble pop!? Should you be afraid!? YES!!!" and also in places like Canada where the economy is predictably crashing because of decades of failures, it's both the cause and answer to macro economic decline. Ironically/suspiciously it's all the same re-hashed redundant takes by everyone from Hank Green to CNBC to every podcast ever, late night shows, radio, everything.
So to me the target of one's annoyance should be the propaganda machine, not the targets of the machine. What are people supposed to feel, totally chill because they have tons of control?
https://www.theguardian.com/society/2025/dec/05/ai-deepfakes...
There is no way Google thinks it's in their interest to serve up clean data to anyone but themselves.
No, gen AI is a subset of machine learning.
Other subfields of AI include things like search, speech and language understanding, knowledge representation, and so on. There’s a lot more to AI than machine learning and a lot more to machine learning than LLMs (“gen AI”).
A lot of folks here hate AI and YouTube and Google and stuff, but it would be more productive to hate them for what they are actually doing.
But most people here are just taking this headline at face value and getting pitchforks out. If you try to watch the makeup guy’s proof, it’s talking about Instagram (not YouTube), doesn’t have clean comparisons, is showing a video someone sent back to him, which probably means it’s a compression artifact, not a face filter that the corporate overlords are hiding from the creator. It is not exactly a smoking gun, especially for a technical crowd.
It's the other way around isn't it? "AI" is a subset of ML.
I wonder if it will end up being treated as part of a codec instead of edits to the base film, and can then be re-run to undo the video's?
It feels like there needs to be a way to verify that what you uploaded is what's on the site.
Strongly recommend. We’ll get local AIs that can skip the cruft soon enough anyway.
I don't understand the justification for the expense or complexity.
And then the discourse is so riddled with misnomers and baited outrage that it goes nowhere.
The other example in submitted post isn't 'edits to videos' but rather the text descriptions of automated captions. The Gemini/AI engine not being very good at summarizing is a different issue.
I just completely despair. What the fuck happened to the internet? Absolutely none of these CEOs give a shit. People need to face real punishments
The key section:
> Rene Ritchie, YouTube’s creator liaison, acknowledged in a post on X that the company was running “a small experiment on select Shorts, using traditional machine learning to clarify, reduce noise and improve overall video clarity—similar to what modern smartphones do when shooting video.”
So the "AI edits" are just a compression algorithm that is not that great.
So counterintuitively, noise reduction improves compression ratios. In fact many video codecs are about determining which portion of the video IS noise that can be discarded, and which bits are visually important...
Or to put it another way, to me it would be similarly disingenuous to describe e.g. dead code elimination or vector path simplification as "just a compression algorithm" because the resultant output is smaller than it would be without. I think part of what has my hackles raised is that it claims to improve video clarity, not to optimise for size. IMO compression algorithms do not and should not make such claims; if an algorithm has the aim (even if secondary) to affect subjective quality, then it has a transformative aspect that requires both disclosure and consent IMO.
I think the first comment is why they would position noise reduction as being both part of their compression and a way to improve video clarity.
It's in the loop of the compression and decompression algorithm.
Video compression has used tricks like this for years. For example, reducing noise before decode and then adding it back in after the decode cycle. Visual noise doesn't need to be precise, so it removing it before compression and then approximating it on the other end saves a lot of bits.
It looks like quality cleanup, but I can’t imagine many creators aren’t using decent camera tech and editing software for shorts.
And as you say, arbitrarily applying quality cleanup is making assumptions of the quality and creative intent of the submitted videos. It would be one thing if creators were uploading raw camera frames to YouTube (which is what smartphone camera apps are receiving as input when shooting video), but applying that to videos that have already been edited/processed and vetted for release is stepping over a line to me. At the very least it should be opt-in (ideally with creators having the ability to preview the output before accepting to publish it).
That being said, I don't believe they should be doing anything like this without the creator's explicit consent. I do personally think there's probably a good use case for machine learning / neural network tech applied to the clean up of low-quality sources (for better transcoding that doesn't accumulate errors & therefore wastes bitrate), in the same way that RTX Video Super Resolution can do some impressive deblocking & upscaling magic[2] on Windows. But clearly they are completely missing the mark with whatever experiment they were running there.
[1] https://www.ynetnews.com/tech-and-digital/article/bj1qbwcklg
[2] compare https://i.imgur.com/U6vzssS.png & https://i.imgur.com/x63o8WQ.jpeg (upscaled 360p)
> "Making AI edits to videos" strikes me as something particularly egregious; it leads a viewer to see a reality that never existed, and that the creator never intended.
YouTube is not applying any "face filters" or anything of the sort. They did however experiment with AI upscaling the entire image which is giving the classic "bad upscale" smeary look.
Like I said, I think that's still bad and they should have never done it without the clear explicit consent of the creator. But that is, IMO, very different and considerably less bad than changing someone's face specifically.
Here's some other creators also talking about it happening in youtube shorts: https://www.reddit.com/r/BeautyGuruChatter/comments/1notyzo/...
another example: https://www.youtube.com/watch?v=tjnQ-s7LW-g
https://www.reddit.com/r/youtube/comments/1mw0tuz/youtube_is...
https://www.bbc.com/future/article/20250822-youtube-is-using...
If you open the context of the comment, they are specifically talking about the bad, entire-image upscaling that gives the entire picture the oily smeary look. NOT face filters.
EDIT : same thing with the two other links you edited into your comment while I was typing my reply.
Again, I'm not defending YouTube for this. But I also don't think they should be accused of doing something they're not doing. Face filters without consent are a far, far worse offense than bad upscaling.
I would like to urge you to be more cautious, and to actually read what you brandish as proof.