New Research in Detecting AI-Generated Videos

The latest in what will be a continuing arms race between creating and detecting videos:

The new tool the research project is unleashing on deepfakes, called “MISLnet”, evolved from years of data derived from detecting fake images and video with tools that spot changes made to digital video or images. These may include the addition or movement of pixels between frames, manipulation of the speed of the clip, or the removal of frames.

Such tools work because a digital camera’s algorithmic processing creates relationships between pixel color values. Those relationships between values are very different in user-generated or images edited with apps like Photoshop.

But because AI-generated videos aren’t produced by a camera capturing a real scene or image, they don’t contain those telltale disparities between pixel values.

The Drexel team’s tools, including MISLnet, learn using a method called a constrained neural network, which can differentiate between normal and unusual values at the sub-pixel level of images or video clips, rather than searching for the common indicators of image manipulation like those mentioned above.

Research paper.

Posted on July 29, 2024 at 7:02 AM10 Comments

Comments

Clive Robinson July 29, 2024 9:00 AM

@ Bruce, ALL,

Re : Demise of an arms race.

“The latest in what will be a continuing arms race”

History shows us that even in a war arms races come to an end.

With the winner generally being the side with,

1, Greatest technological capacity.
2, Largest reserves of resources.

Because of two reasons,

The first is an arms race gets expensive to some power law for each increment in capability.

The second is some technology obsoletes the technology the arms race is based on, and effectively a new arms race begins.

Perhaps not as well known as it should be but LLM AI systems in their current form have got beyond the size where the laws of physics and basic geometry allow simple scaling up, because whilst much can be done in parallel not every thing can and it takes quite a while for a signal to get from one corner of a cube to the furthest corner.

A grain of sand is considered about 1pS or picosecond in size and a one foot length 1nS or nano-second. As these are the times it takes light in freespace to travel these distances.

But there is also a density issue. Such electronics draws a European sized countries worth of electricity and in terms of power in to power out current LLMs are about as inefficient as it gets. Which means that most of that input power becomes heat. Air is not a particularly good conductor of heat and even though water is better it has a lot of disadvantages. Copper silver, gold and diamond do a much better job as conductors of heat but there are other limits on their use.

So even with optimal step down techniques to move the heat away, only about 1/10,000th of any given volume can be considered as being computing volume often less a lot less.

I could give other reasons for why current LLM AI systems are hitting the cost and resource brick walls of physics but that’s not really that interesting. We know it’s going to happen and realistically when even with significant algorithmic improvements.

What is interesting is the second way Arms Races end, that is how the current LLM systems Arms Race will be replaced with a new AI arms race. For the cycle to start all again.

Oh and at some point when things move to Carbon and potentially Iron based systems as Silicon and Aluminium reach their limits.

Clive Robinson July 29, 2024 9:32 AM

@ Bruce, ALL,

Re : Is it plagiarism or coincidence?

I suspect some of you remember that a little while ago I pointed out a certain “naughty wordlist” member was a term of art in a knowledge domain.

And that it was a more correct if not appropriate word to use than “Hallucinate” for LLM AI systems churning out nonsense…

Well it looks like others are picking up on this and on 25th July just last week put up,

https://www.livescience.com/technology/artificial-intelligence/ai-isnt-hallucinating-its-bullshitting

It’s basically what I’d said “fleshed out” because they had a little more space to fill 😉

Anyway there is an “original paper / publication” at,

https://link.springer.com/article/10.1007/s10676-024-09775-5

That I amplified a little.

For various reasons I suggest the use of “the correct term of art”. As it might help the general public get their head around things. After all most humans BS a lot as a form of humour or for comical effect and it’s seen as “a human thing to do”. Whereas few humans hallucinate without their being some aberrant cause. Thus it’s seen as something dark and dangerous for which people get locked away for.

Winter July 29, 2024 10:05 AM

The latest in what will be a continuing arms race between creating and detecting videos:

Like every arms race, the limits on this arms race will be set by resource limitations. Especially, the difficulty of collecting training data and the cost of training ever larger models on ever larger training sets will become a problem.

Obviously, people will have to “work smarter, not harder” 😉

Gert-Jan July 30, 2024 6:32 AM

In the long run, I think it’s impossible for detection software to determine that something is AI generated.

But that’s a bit besides the point. First of all, such detection system can only guarantee (or least indicate with high degree of certainty) that the work has been manipulated. It can’t guarantee that something was created without any such manipulation (false negatives).

Research has shown that fact checking social media posts is a process that does not work very well to prevent the spread is misinformation and disinformation.

Something similar applies here. There will always be a delay between something from a malicious source hitting the public eye and the time it gets classified as a “fake”.

Not to forget, that when it’s about a polarizing issue, the classification as “fake” will be called into question or even outright rejected.

Yes, it’s a tool in the toolbox. But it won’t win the war.

Clive Robinson July 30, 2024 8:57 AM

@ Winter,

Re : Resource limitations.

“Like every arms race, the limits on this arms race will be set by resource limitations.”

But which ones will bite first?

There are limitations such as the speed of light and thermodynamics that there is no known way to beat and quite a few LLM systems have got quite close to those.

Then there is the density issues of getting energy in and heat out hence resources like power, water, and metals like copper, silver and gold are needed in high quantities.

Another issue is silicon and heat getting silicon semiconductor devices above 250C is at best problematic. Which is why some are looking at carbon and iron alloys.

But then there are “Human Failings”.

Yup call it greed or some other “sin” but it’s come to light that four organisations have been stockpiling those Nvidia H100 chips that cost upto $40,000 each.

And we are not talking little piles we are talking getting on for the better part of $100,000,000,000 tucked away out of sight.

One of those four organisations is a large Venture Capitalist. And they are apparently “renting out” their H100 systems to “Start Ups”, but not for money, but A stock equity…

But it’s interesting to note that these high end systems for LLMs show all the disadvantages that “block chain” systems did…

Winter July 31, 2024 1:31 AM

@Clive

Re : Resource limitations.

But which ones will bite first?

Look at biological brains.

Brains operate >10^15 synapses with only tens of Watts. The energy/thermodynamics limitation is currently been improved by using brain inspired neural spikes instead of synchronized activation “levels”.
‘https://www.nature.com/articles/s41467-024-47811-6.pdf

Neuromorphic computing with spikes has already been realized with ~2pJ per Synapse operation.
‘https://dl.acm.org/doi/fullHtml/10.1145/3588591

Biology shows that many AI tasks can be learned with “small” amounts of data. A 4 year old is estimated to have heard ~20-50 million words, but can already communicate well.
‘https://laulima.hawaii.edu/access/content/user/cawdery/ED_282/EARLY_CATASTROPHE_ED_297.pdf

Compare that with the hundreds of billions of words necessary for LLMs and there is ample room for improvement in the data efficiency of AI training. It is just that we have no idea how to make current training more efficient.

So, there are currently known routes to reduce energy use, but there is no plausibel route known yet to increase data efficiency.

Clive Robinson August 1, 2024 6:58 AM

@ Winter,

First off mixing units makes mental comparisons harder than it needs to be when trying to follow an argument.

Thus “watts” and “1pJ” are not directly comparable due to time.

But also we know that the biological neural nets function very differently as far as usage is concerned.

A current AI LLM is run effectively as a single network with all parts functioning.

Fast MRI scanning and similar show the brain functions as many almost independent networks that “ramp up and ramp down” their usage with what we believe is function (though might be something else).

Interestingly we know that whilst the output of a biological neuron might be linear a subsequent neuron receiving two or more such linear signals will not see the combination linearly.

These are not things current digital neural nets emulate at all, let alone on a spectrum of approximately to closely.

Are these paths worth investigating?

Yes because they have significant advantages in reducing amongst other things power density issues.

But aside from a few computer simulations little has been done on this, and realistically doing so would to many appear to be turning the AI clock back two or more decades.

Winter August 2, 2024 2:07 AM

@Clive

Thus “watts” and “1pJ” are not directly comparable due to time.

Obviously, but as spike based computing scales with synapse processing and spikes, energy per synapse switch is a more relevant measure of energy consumption than energy per time.

Basically, when a conventional CPU/GPU is idle, they still consume a lot of power, but a spike based system will “run” at nearly no power when there are no spikes. The power use profiles of conventional GPUs and neuromorphic chips are not easily comparable.

These are not things current digital neural nets emulate at all, let alone on a spectrum of approximately to closely.

Spike based neuromorphic networks are designed to emulate all these functions. As the links I supplied show, there exist functioning spike based networks but they are still in the early phases. They do show that dramatic reductions in energy use are feasible, and expected.

‘https://www.nature.com/articles/s41467-024-47811-6.pdf

‘https://dl.acm.org/doi/fullHtml/10.1145/3588591

ResearcherZero August 2, 2024 3:48 AM

How prevalent is AI misinformation?

‘https://www.nature.com/articles/d41586-024-01588-2

Information Voids

A study in Nature last month highlights a previously underappreciated aspect of this phenomenon: the existence of data voids, information spaces that lack evidence, into which people searching to check the accuracy of controversial topics can easily fall. The paper suggests that media-literacy campaigns that emphasize ‘just searching’ for information online need to become smarter.

‘https://www.nature.com/articles/d41586-024-00030-x

“Indeed, genuine, lasting solutions to a problem that could be existential for democracies needs to be a partnership between search-engine providers and sources of evidence-based knowledge.”

Online searches to evaluate misinformation can increase its perceived veracity
https://www.nature.com/articles/s41586-023-06883-y

The fallout and it’s damage…

‘https://www.wired.com/story/qanon-destroys-american-families/

A conceptual framework to characterize points of infodemic intervention.

‘https://pandemicresponse.columbia.edu/what-were-the-information-voids-analyzing-3086-questions-asked-by-dear-pandemic-readers/

Addressing the gaps in existing infodemic management frameworks.

“infodemics are not only driven by misinformation and disinformation; rather, they can influence and be influenced by the broader information ecosystem, which refers to the dynamics of how people consume, produce, interact with, and behave around information.”

During emergencies such as a pandemic, both the supply and demand sides of the information ecosystem can undergo rapid changes.
https://www.thelancet.com/journals/lanpub/article/PIIS2468-2667(24)00031-8/fulltext

Fighting the infodemic: the 4 i Framework for Advancing Communication and Trust
https://bmcpublichealth.biomedcentral.com/articles/10.1186/s12889-023-16612-9

Clive Robinson August 5, 2024 11:32 AM

@ Bruce, ALL,

Apparently some AI Video’s are not as popular as their creators would hope,

https://arstechnica.com/culture/2024/08/google-pulls-its-terrible-pro-ai-dear-sydney-ad-after-backlash/

Apparently Google, put a video advert together especially for the Olympics…

And people made serious complaint and now Google have pulled it… Just like Apple did a little while before, with their squeezy advert.

If AI is as hard a sale to ordinary consumers as it’s apparently currently going… Then those trillion dollar “investment opportunities” may not even be worth “Fire Sale” “scrap value” in a few months.

Leave a comment

Blog moderation policy

Login

Allowed HTML <a href="URL"> • <em> <cite> <i> • <strong> <b> • <sub> <sup> • <ul> <ol> <li> • <blockquote> <pre> Markdown Extra syntax via https://michelf.ca/projects/php-markdown/extra/

Sidebar photo of Bruce Schneier by Joe MacInnis.