Microsoft joins coalition to scrub revenge and deepfake porn from Bing

Microsoft announced it has partnered with StopNCII to help remove non-consensual intimate images — including deepfakes — from its Bing search engine.

When a victim opens a “case” with StopNCII, the database creates a digital fingerprint, also called a “hash,” of an intimate image or video stored on that individual’s device without their needing to upload the file. The hash is then sent to participating industry partners, who can seek out matches for the original and remove them from their platform if it breaks their content policies. The process also applies to AI-generated deepfakes

→ Continue reading at Engadget

Similar Articles

Advertisment

Most Popular