2023 Transparency Report (First Half)

For Pornhub, nothing is more important than the safety of our community. Upholding our Core Values of consent, freedom of sexual expression, authenticity, originality, and diversity is only possible through the continual efforts to ensure the safety of our users and our platform. Consequently, we have always been committed to preventing and eliminating illegal content, including non-consensual material and child sexual abuse material (CSAM). Every online platform has the moral responsibility to join this fight, and it requires collective action and constant vigilance.

Over the years, we have put in place robust measures to protect our platform from abuse. We are constantly improving our Trust and Safety policies to better identify, flag, remove, review, and report illegal or abusive material. We are committed to doing all that we can to ensure a safe online experience for our users, which involves a concerted effort to innovate and always striving to do more.

As part of our efforts and dedication to Trust and Safety, we feel it is important to provide information about our Trust and Safety initiatives, technologies, moderation practices and how our platform responds to harmful content. This Transparency Report provides these insights and builds upon our previous reports. Further information can be found in other sections of our Trust and Safety Center, including resources about our community, keeping our platform safe, and our policies.

We are committed to being as transparent as possible and to being a leader in eradicating abusive content from the internet. We will continue to do our part to create a safe space for our users.

Trust and Safety Initiatives and Partnerships - 2023 Mid-Year Update

Our goal to achieve and set the highest level of Trust and Safety standards relies heavily on not only our internal policies and moderation team, but also on our array of external partnerships and technology. The organizations we partner with are part of the global effort to reduce illegal material online and help ensure our platform and adult users can feel safe.

We continued to further our Trust and Safety efforts and are proud to have added new partnerships and initiatives. Below is a non-exhaustive list of partnerships and accomplishments achieved in the first half of 2023:

IWF Partnership

We continue to participate in our two-year pilot project in partnership with the Internet Watch Foundation (IWF) to develop a model of good practice to support the adult industry in combatting child sexual abuse material online.

The project is being overseen by an external advisory board of experts who will guide the outputs of the partnership. Key learnings from the project will be shared publicly to support others’ work in this area.

Trust and Safety Professional Association

In 2023, we became members of the Trust and Safety Professional Association (TSPA), an organization which works to serve as a global community for Trust and Safety teams and individuals. TSPA offers Trust and Safety teams the opportunity to collaborate and share information to assist in developing best practices related to Trust and Safety on online platforms.

Take It Down Service

We are a participant of the Take It Down service, which is operated by the National Center for Missing and Exploited Children (NCMEC). This service allows individuals to fingerprint their own underage content to assist in the removal and prevention of this content from appearing on online platforms.

StopNCII

Pornhub is an industry partner of StopNCII.org, which allows adults (18+) experiencing Non-Consensual Intimate Image (NCII) abuse to create digital fingerprints of their images and videos to help prevent content being shared across participating platforms. Using groundbreaking hashing technology, StopNCII.org ensures that the original image or video never leaves the user’s device. This service is operated by SWGfL who run the Revenge Porn Helpline, an organization that supports individuals in the UK experiencing intimate image abuse.  

For a full list of Trust and Safety partnerships, initiatives and external technology, please refer to our Trust and Safety Initiatives page.

Content Removals

A total of 1.9 million (1,210,971 videos and 706,539 photos) pieces of content were uploaded to Pornhub in 2023 from January through June. Of these uploads, 95,822 (65,900 videos and 29,922 photos) were blocked at upload or removed for violating our Terms of Service.

An additional 116,246 pieces of content (100,808 videos and 15,438 photos) uploaded prior to 2023 were labeled as Terms of Service violations, for a total of 212,068 pieces of content (166,708 videos and 45,360 photos) removed during the six month period ended June 30, 2023 for violations of our Terms of Service.

Picture1.png

The above figure provides a breakdown of content uploaded and removed in the first six months of 2023 by removal reason. For additional information and explanations of each removal reason, please see below.

Non-Consensual Content (NCC) 

In the first half of 2023, we removed 4,449 pieces of content (3,232 videos and 1,217 photos) for violating our NCC policy. When we identify or are made aware of the existence of non-consensual content on Pornhub, we remove the content, fingerprint it to prevent re-upload, and ban the user in accordance with our Terms of Service

Our NCC policy includes the following violations:

  • Non-Consensual Content – Act(s): Content which depicts activity for which an individual did not validly consent. A non-consensual act is one involving the application of force directly or indirectly without the consent of the other person to a sexual act. 
  • Non-Consensual Content – Recording(s): The recording of sexually explicit activity without the consent of the featured person. 
  • Non-Consensual Content – Distribution: The distribution of sexually explicit content without the featured person’s permission. 
  • Non-Consensual Content – Manipulation: Also known as “deepfake,” this type of content appropriates a real person’s likeness without their consent. 

Potential Child Sexual Abuse Material (CSAM) 

We have a zero-tolerance policy toward any content that features a person (real or animated) that is under the age of 18. Through our use of cutting-edge CSAM detection technology, identity verification measures, as well as human moderators, we strive to quickly identify and remove potential CSAM and ban users who have uploaded this content. Any content that we identify as or are made aware of being potential CSAM is also fingerprinted to prevent re-upload upon removal. We report all instances of potential CSAM to the National Center for Missing and Exploited Children (NCMEC).

In the first six months of 2023, we made 1,214 reports to NCMEC. These reports contained 3,951 pieces of content (2,632 videos and 1,319 photos) which were reported for violating our CSAM Policy.

Community Guidelines Violations

In the first half of 2023, we removed 80,021 pieces of content (40,007 videos and 40,014 photos) for violating our Community Guidelines. When we identify or are made aware of the existence of content that violates our Community Guidelines on Pornhub, we remove the content, fingerprint it to prevent re-upload, and, where appropriate, may ban the user. 

A portion of these removals is related to requests from individuals who previously consented to the distribution of content they appear in but have submitted a request to retract that consent. Uploaders whose content is removed upon an individual retracting their consent for the content to be distributed are not subject to banning as per our NCC policy. Individuals who wish to retract their consent may reach out to us via our Content Removal Request form.

Content may be removed for Community Guidelines violations, outside of any of the other reasons listed, for being objectionable or inappropriate for our platform. This also includes violations related to brand safety, defamation, and the use of our platform for sexual services.

Other Reasons for Removal

Hate Speech

We do not allow abusive or hateful content or comments of a racist, homophobic, transphobic, or otherwise derogatory nature intended to hurt or shame those individuals featured.

From January through June, we removed 34 pieces of content (all videos) for violating our hate speech and violent speech policy.   

Animal Guidelines

Content that shows or promotes harmful activity to animals, including exploiting an animal for sexual gratification, physical or psychological abuse of animals, is prohibited on our platform.

In the first six months of 2023, we removed 197 pieces of content (187 videos and 10 photos) for violating our Animal Guidelines.   

Bodily Fluids

Content depicting vomit, or feces, among others, is not permitted on Pornhub.

In the first half of 2023, we removed 1,691 pieces of content (1,660 videos and 31 photos) for violating our policy related to excretion of bodily fluids.   

Familial Relations

We do not allow any content which depicts incest.

During the six month period ended June 30, 2023, we removed 1,533 pieces of content (1,532 videos and 1 photo) for violating our policy related to familial relations.  

Bodily Harm

We prohibit excessive use of force that poses a risk of serious physical harm to any person appearing in the content or any physical or sexual activity with a corpse. This violation does not include forced sexual acts, as this content would fall under our non-consensual policy. 

In the first six months of 2023, we removed 545 pieces of content (530 videos and 15 photos) for violating our bodily harm and violent content policy.   

Co-Performer Verification Incomplete

We require uploaders to maintain valid documentation for any other individuals appearing in their content. We may remove content if verification for a co-performer is incomplete.

In the first half of 2023, we removed 119,647 pieces of content (116,894 videos and 2,753 photos) for having incomplete verification of the co-performers depicted.  

Content Uploaded in January Through June, 2023

An ongoing goal of our moderation efforts is to identify as much unwanted content as possible at the time of upload before appearing live on Pornhub.

Videos uploaded in the first half of 2023 that were removed for violating our Terms of Service accounted for 0.71% of total views of videos uploaded during the six month period ended June 30, 2023.

Picture2.png

The above figure shows the number of times videos uploaded between January and June, 2023 were viewed prior to their removal for Terms of Service Violations.

Analysis: More than three-quarters of removed videos were identified and removed by our human moderators or internal detection tools before being viewed on our platform.

Moderation and Content Removal Process

Upload and Moderation Process

As Pornhub strives to prevent abuse on our platform, all content is reviewed upon upload by our trained staff of moderators before it is ever made live on our site. Uploads are also automatically scanned against several external automated detection technologies as well as our internal tool for fingerprinting and identifying violative content, Safeguard. Furthermore, only verified users can upload content to our platform, after passing verification measures through Yoti.

Picture3.png

Content Removal Process

Users may alert us to content that may be illegal or otherwise violates our Terms of Service by using the flagging feature or by filling out our Content Removal Request form. 

While users can notify us of potentially abusive content, we also invest a great deal into internal detection, which includes technology and a 24/7 team of moderators, who can detect content that may violate our Terms of Service at time of upload and before it is published. Most of this unwanted content is discovered via our internal detection methods.

Picture4.png

The above figure shows a breakdown of removals based on the source of detection. In addition to internal and user-sourced removals, removals due to External Trusted Sources include legal requests and notifications from our Trusted Flaggers.

Analysis: Of all the content that was removed for a TOS violation in the first six months of 2023, nearly 90% of content was identified via Pornhub’s internal methods of moderation, compared to 9.86% of content being alerted to our team from our users. 

As we add technologies and new moderation practices, our goal is to ensure that our moderation team is identifying as much violative content as possible before it is published. This is paramount to combatting abusive material and for creating a safe environment for our users. From content uploaded during the first six months of 2023, 92.23% of removals were the result of Pornhub’s internal detection methods. 

Picture5.png

The above figure shows a breakdown of removals that were uploaded in the first half of 2023 based on the source of detection. See above for further information about the sources of removal.  

Analysis: Of all the content that was uploaded and removed for a TOS violation during the first six months of 2023, 92.23% of content was identified via Pornhub’s internal methods of moderation, compared to 7.69% of content being alerted to our team from our users. 

User-Initiated Removals

While we endeavor to remove offending content before it is ever available on our platform, our users play an important role in alerting us to content that may violate our Terms of Service. Users may bring this content to our attention by flagging content or submitting a content removal request. These features are available to prevent abuse on the platform, assist victims in quickly removing content, and to foster a safe environment for our adult community.

Users may alert us to content that may be illegal or otherwise violate our Terms of Service by using the flagging feature found below all videos and photos, or by filling out our Content Removal Request form.  Flagging and Content Removal Requests are kept confidential, and we review all content that is brought to our attention by users through these means.

User-based removals accounted for 7.69% of all removals for content uploaded and removed in the first six months of 2023. After internal detection, user-based removals account for the largest percentage of removals overall, and specifically when identifying non-consensual content.

User Flags

Users may flag content to bring it to the attention of our moderation team for the following reasons: Potentially Features a Minor, Violent or Harmful Acts, Hateful or Inflammatory, Spam or Misleading Content or Otherwise Objectionable or Inappropriate.

From January through June, content reported using the flagging feature was most often for “Violent or Harmful Acts” and “Otherwise Inappropriate or Objectionable”, which accounts for nearly 60% of flags.

Picture6.png

The figure above shows the breakdown of flagged videos received from users during the six month period ended June 30, 2023.

Upon receiving flags on videos and other content, our moderation team reviews the content and removes it if it violates our Terms of Service. Of the content removed after being flagged, most were removed due to Community Guidelines violations.

In addition to flagging content, users may also flag comments, users, and verified uploaders.

Picture7.png

The figure above shows a breakdown of reasons of why a video was removed after it had been flagged by a user. After review by our moderation team, most of the content removed was due to Community Guidelines violations.

Analysis: In the figure titled, “Videos Flagged by Reason”, the second largest category selected by users was “Otherwise Inappropriate or Objectionable”, which correlates to Community Guidelines violation. Community Guidelines violations accounted for over 70% of flagged videos that were removed.

Content Removal Requests

Upon receipt of a Content Removal Request, the content is immediately suspended from public view pending investigation. This is an important step in assisting victims and addressing abusive content. All requests are reviewed by our moderation team, who work to ensure that any content that violates our Terms of Service is quickly and permanently removed. In addition to being removed, abusive content is digitally fingerprinted with both Mediawise and Safeguard, to prevent re-upload. In accordance with our policies, uploaders of non-consensual content or potential CSAM will also be banned. All content identified as potential CSAM is reported to the National Center for Missing and Exploited Children (NCMEC).

During the first six months of 2023, Pornhub received a total of 6,888 Content Removal Requests.

Picture8.png

The figure above shows the number of Content Removal Requests received by Pornhub per month in the first half of 2023.

Of the content removed due to Content Removal Requests, the majority was removed for Withdrawal of Consent. 

Picture9.png

The figure above displays the breakdown of removal reasons for content that was removed due to Content Removal Requests.

Analysis: 17.26% of content that was removed due to a Content Removal Request was due to non-consensual content, which is the third most common reason for removal after Withdrawal of Consent and Community Guidelines Violations.

Combatting Child Sexual Abuse Material

We have a firm zero-tolerance policy against underage content, as well as the exploitation of minors in any capacity.

Pornhub is against the creation and dissemination of child sexual abuse material (CSAM), as well as the abuse and trafficking of children. CSAM is one of many examples of illegal content that is prohibited by our Terms of Service and related guidelines.

Pornhub is proud to be one of the 1,400+ companies registered with the National Center for Missing and Exploited Children (NCMEC) Electronic Service Provider (ESP) Program. We have made it a point to not only ensure our policies and moderation practices are aligned with global expectations on child exploitation, but also to publicly announce our commitment and support to continue to adhere to the Voluntary Principles to Counter Online Child Sexual Exploitation and Abuse.

Potential child sexual abuse material that is detected through moderation efforts, including detection through various internal and external tools listed here, or reported to us by our users and other external sources is reported to NCMEC. For the purposes of our CSAM Policy and our reporting to NCMEC, we deem a child to be a human, real or fictional, who is or appears to be below the age of 18. Any content that features a child engaged in sexually explicit conduct, lascivious or lewd exhibition is deemed CSAM for the purposes of our Child Sexual Abuse Material Policy.

Reported Potential CSAM

In the first six months of 2023, we made 1,214 reports to the National Center for Missing and Exploited Children (NCMEC). These reports contained 3,951 pieces of content (2,632 videos and 1,319 photos) that were identified as potential CSAM.

Internal detection was the greatest source of identifying potential CSAM that was removed in the first half of 2023.

Picture10.png

The above figure shows a breakdown of reported potential CSAM based on the source of detection. In addition to internal and user-sourced removals, removals due to External Trusted Sources include legal requests and notifications from our Trusted Flaggers.

Analysis: Of all the content that was reported as potential CSAM during the six month period ended June 30, 2023, including content uploaded prior to 2023, nearly 80% of content was identified via Pornhub’s internal methods of moderation, compared to 14.33% of content being alerted to our team from our users. 

For content uploaded between January through June that was reported as potential CSAM, 81.93% was identified through our internal detection methods.

Picture11.png

The above figure shows a breakdown of reported potential CSAM based on the source of detection. In addition to internal and user-sourced removals, removals due to External Trusted Sources include legal requests and notifications from our Trusted Flaggers.

Analysis: Of the content that was both uploaded and reported as potential CSAM in the first half of 2023, 81.93% of content was identified via Pornhub’s internal methods of moderation, compared to the remainder of content being alerted to our team from our users. None of this content was removed due to notifications from External Trusted Sources.   

As part of our efforts to identify and quickly remove potential CSAM and other harmful material, we strive to ensure this content is removed before it is available to users and consumers. Of the videos uploaded and reported as potential CSAM in the first half of 2023, which represents 0.07% of videos uploaded from January through June, 69.33% were removed before being viewed.   

Videos uploaded in the first six months of 2023 and removed for violating our CSAM policy accounted for 0.01% of total views of all videos uploaded in the first six months of 2023. Picture12.png

The above figure shows the number of times videos uploaded in the first six months of 2023 were viewed prior to being reported as potential CSAM.

Analysis: Nearly 70% of reported videos were identified and removed by our human moderators or internal detection tools before being viewed on our platform.

Non-Consensual Content

Pornhub is an adult content-hosting and sharing platform for consenting adult use only. We have a zero-tolerance policy against non-consensual content (NCC). 

Non-consensual content not only encompasses what is commonly referred to as “revenge pornography” or image-based abuse, but also includes any content which involves non-consensual act(s), the recording of sexual material without consent of the person(s) featured, or any content which uses a person’s likeness without their consent.  

We are steadfast in our commitment to protecting the safety of our users and the integrity of our platform; we stand with all victims of non-consensual content.

From January through June, we removed 4,449 pieces of content (3,232 videos and 1,217 photos) for violating our NCC policy.  

For all content removed for violating our NCC policy in the first six months of 2023, the majority of this content was identified through users, at 79.34%.

Picture13.png

The above figure shows a breakdown of NCC removals based on the source of detection. In addition to internal and user-sourced removals, removals due to External Trusted Sources include legal requests and notifications from our Trusted Flaggers.

Analysis: Of all the content that was removed for violating our NCC policy in the first six months of 2023, including content uploaded prior to 2023, 17.76% of content was identified via Pornhub’s internal methods of moderation, compared to 79.34% of content being alerted to our team from our users. 

Most of the content uploaded in the first half of 2023 and removed for a violation of our NCC policy was identified through users, at 73.03%.

Picture14.png

The above figure shows a breakdown of NCC removals that were uploaded during the six month period ended June 30, 2023, based on the source of detection. See above for further information about the sources of removal.

Analysis: Of all the content that was uploaded and removed for violating our NCC policy during the first six months of 2023, 25.82% of content was identified via Pornhub’s internal methods of moderation, compared to nearly three quarters of content being alerted to our team from our users.

As part of our efforts to identify and quickly remove non-consensual content and other harmful material, we try to ensure this content is removed before it is available to users and consumers. Of videos both uploaded and removed for violating our NCC policy in 2023 from January through June, which represents 0.09% of uploads during this time, nearly 25% were removed before being viewed. 

Videos uploaded in the first six months of 2023 and removed for violating our NCC policy accounted for 0.07% of total views of all videos uploaded in the first half of 2023.

Picture15.png

The above figure shows the number of times videos uploaded in the first six months of 2023 were viewed prior to being removed for violating our NCC policy.

Analysis: 24.81% of videos were identified and removed by our human moderators or internal detection tools before going live on our platform.

Banned Accounts

As part of our efforts to ensure user and platform safety, we ban accounts that have committed serious or repeated violations of our Terms of Service. Additionally, accounts that have violated our CSAM Policy are reported to the National Center for Missing and Exploited Children (NCMEC).

In the first six months of 2023, we banned a total of 19,852 user accounts for violating our Terms of Service.  This includes accounts that were banned for uploading content that was removed, as well as accounts that were banned for violations not related to content, such as spamming. Additionally, an account may be banned if we identify or are made aware that the account belongs to an individual whose previous account was banned for violating our Terms of Service.  

Picture16.png

The above figure shows the breakdown by ban reason for accounts that were banned in 2023 from January through June.  

Analysis: 70.37% of accounts were banned for a Community Guidelines violation, while nearly 10% of accounts were banned as they were identified as belonging to an individual whose previous account(s) were banned.  

The majority of accounts banned during this period were identified by our internal detection methods.

Picture17.png

The above figure shows a breakdown of banned accounts based on the source of detection. In addition to internal and user-sourced removals, removals due to External Trusted Sources include legal requests and notifications from our Trusted Flaggers.

Analysis: Of the accounts that were banned in the first six months of 2023, nearly 90% were identified due to internal detection methods, while user notifications accounted for 11.12% of banned accounts.  

DMCA

The Digital Millennium Copyright Protection Act, known as the DMCA, is a U.S. law that limits the liability of online service providers for the copyright infringement caused by their users. To qualify, online service providers must do —and not do — certain things. The most well-known aspects of the DMCA are its requirements that a service provider remove materials after it receives a notice of infringement and have (and implement) a policy to deter repeat infringement. Pornhub promptly responds to requests relating to copyright notifications.

In addition to blocking uploads of works previously fingerprinted with Mediawise or Safeguard from appearing on Pornhub, during the six month period ended June 30, 2023, we responded to 3,868 requests for removal, leading to a total of 2,720 pieces of content removed.

Picture18.png

The above figure shows the number of requests for removal related to DMCA per month in the first half of 2023.

Cooperation with Law Enforcement

We cooperate with law enforcement globally and readily provide all information available to us upon request and receipt of appropriate documentation and authorization. Legal requests can be submitted by governments and law enforcement, as well as private parties.

In the first half of 2023, we responded to 125 legal requests. A breakdown by request type as indicated by law enforcement is provided below. 

Picture19.png

The above figure displays the number of requests Pornhub received from law enforcement and legal entities in the first six months of 2023 broken down by the type of request as indicated by the requester.

As a result of these requests, we classified a total of 14 pieces of content (all videos) as Terms of Service Violations due to legal requests. A breakdown by removal type is provided below. 

Picture20.png

The above figure displays photos and videos that were identified in law enforcement requests that we deemed to violate our Terms of Service upon review. 

Content De-Indexing

As part of our Trust and Safety efforts to remove abusive content not only from our platform, but globally, we request de-indexing of URLs that were removed for violating our Terms of Service from search engines. We proactively seek to de-index all content that is removed for potential CSAM and non-consensual content. In 2023, we added the proactive process for de-indexing content removed for Withdrawal of Consent.

In the first half of 2023, we requested de-indexing of 26,570 URLs for 20,986 unique pieces of content.

Picture21.png

The above figure displays a breakdown of the reason for URLs that were de-indexed from search engines. There may be several URLs de-indexed per piece of content, as we request the de-indexing of every version of the URL (e.g. it.pornhub.com/123 would be a URL de-indexed for the content located at www.pornhub.com/123).

Was this article helpful?
30 out of 34 found this helpful