TNSA Legal Center

Policies, terms, and compliance information for TNSA. Browse categories or jump to a specific topic.

TNSA Content Moderation & Safety Framework

Comprehensive policies governing content restrictions, safety measures, enforcement procedures, and appeals processes across all TNSA AI services. Version 4.1 — Effective December 9, 2025

1. CONTENT MODERATION PRINCIPLES

1.1 Core Safety Principles

  • 1.1.1 TNSA is committed to maintaining safe, respectful, and lawful use of all AI services and platforms.
  • 1.1.2 Content moderation decisions are based on objective criteria, community standards, and applicable laws.
  • 1.1.3 We prioritize user safety while respecting freedom of expression and legitimate use cases.
  • 1.1.4 Moderation policies apply equally to all users regardless of status, affiliation, or subscription level.
  • 1.1.5 We employ both automated systems and human reviewers to ensure comprehensive content oversight.
  • 1.1.6 Transparency and accountability guide our moderation practices and policy development.
  • 1.1.7 We continuously improve our systems based on user feedback and emerging safety research.
  • 1.1.8 Cultural sensitivity and global perspectives inform our content evaluation processes.
  • 1.1.9 We collaborate with industry partners, researchers, and civil society to enhance safety measures.
  • 1.1.10 Regular audits and assessments ensure the effectiveness of our moderation framework.

1.2 Scope of Moderation

  • 1.2.1 Content moderation applies to all user inputs, generated outputs, and platform interactions.
  • 1.2.2 Moderation covers text, images, audio, video, and multimodal content across all TNSA services.
  • 1.2.3 Both real-time and post-processing moderation techniques are employed as appropriate.
  • 1.2.4 API usage, web interfaces, and mobile applications are subject to consistent moderation standards.
  • 1.2.5 User-generated content, AI-generated outputs, and hybrid content are all within scope.
  • 1.2.6 Private conversations, public posts, and shared content receive appropriate levels of oversight.
  • 1.2.7 Metadata, file names, and associated information may be subject to moderation review.
  • 1.2.8 Cross-platform content sharing and external integrations are monitored for policy compliance.
  • 1.2.9 Bulk operations, automated submissions, and high-volume usage receive enhanced scrutiny.
  • 1.2.10 Research, educational, and journalistic content may receive specialized review processes.

1.3 Moderation Standards

  • 1.3.1 Content policies are designed to prevent harm while enabling legitimate creative and professional use.
  • 1.3.2 Context, intent, and potential impact are considered in all moderation decisions.
  • 1.3.3 Age-appropriate content standards are enforced based on service terms and local regulations.
  • 1.3.4 Professional, educational, and artistic content may receive different evaluation criteria.
  • 1.3.5 Satirical, fictional, and clearly labeled creative content is distinguished from factual claims.
  • 1.3.6 Historical, documentary, and news-related content receives specialized consideration.
  • 1.3.7 Medical, legal, and financial advice is subject to enhanced accuracy and safety requirements.
  • 1.3.8 Technical documentation, code, and instructional content follows domain-specific guidelines.
  • 1.3.9 Multilingual content is evaluated by culturally competent reviewers when possible.
  • 1.3.10 Emerging content types and use cases are evaluated against core safety principles.

2. PROHIBITED CONTENT CATEGORIES

2.1 Illegal Content

  • 2.1.1 Content that violates applicable laws in the user's jurisdiction or TNSA's operating jurisdictions.
  • 2.1.2 Child sexual abuse material (CSAM) or content that sexualizes, grooms, or otherwise harms children.
  • 2.1.3 Non-consensual intimate imagery, including deepfakes and synthetic media of real persons.
  • 2.1.4 Content that facilitates human trafficking, exploitation, or modern slavery.
  • 2.1.5 Instructions for creating illegal weapons, explosives, or dangerous substances.
  • 2.1.6 Content that promotes or facilitates illegal drug manufacturing or distribution.
  • 2.1.7 Fraudulent schemes, scams, or content designed to deceive for financial gain.
  • 2.1.8 Copyright infringement, trademark violations, or other intellectual property theft.
  • 2.1.9 Content that violates export control laws or sanctions regulations.
  • 2.1.10 Money laundering, terrorist financing, or other financial crimes facilitation.

2.2 Harmful Content

  • 2.2.1 Content that promotes, encourages, or provides instructions for self-harm or suicide.
  • 2.2.2 Graphic violence, torture, or content designed to shock or disturb viewers.
  • 2.2.3 Hate speech targeting individuals or groups based on protected characteristics.
  • 2.2.4 Harassment, bullying, doxxing, or content intended to intimidate or threaten.
  • 2.2.5 Misinformation that could cause imminent physical harm or public safety risks.
  • 2.2.6 Content that promotes dangerous activities, challenges, or risky behaviors.
  • 2.2.7 Eating disorder promotion, pro-anorexia, or content glorifying unhealthy body image.
  • 2.2.8 Content that could facilitate stalking, surveillance, or privacy violations.
  • 2.2.9 Malicious code, viruses, or content designed to compromise computer systems.
  • 2.2.10 Content that promotes or facilitates addiction to harmful substances or behaviors.

2.3 Deceptive Content

  • 2.3.1 Deepfakes or synthetic media designed to deceive about real events or persons.
  • 2.3.2 Impersonation of public figures, organizations, or other individuals without disclosure.
  • 2.3.3 False claims about medical treatments, cures, or health interventions.
  • 2.3.4 Misleading information about elections, voting processes, or democratic institutions.
  • 2.3.5 Conspiracy theories that could lead to real-world harm or violence.
  • 2.3.6 False emergency alerts or crisis information that could cause panic.
  • 2.3.7 Manipulated media presented as authentic without appropriate labeling.
  • 2.3.8 Phishing attempts, social engineering, or identity theft facilitation.
  • 2.3.9 False academic credentials, certifications, or professional qualifications.
  • 2.3.10 Misleading financial advice or investment schemes designed to defraud.

2.4 Privacy Violations

  • 2.4.1 Sharing of private personal information without consent (doxxing).
  • 2.4.2 Non-consensual sharing of private communications or confidential information.
  • 2.4.3 Content that facilitates stalking, surveillance, or unwanted contact.
  • 2.4.4 Biometric data processing without appropriate consent and safeguards.
  • 2.4.5 Location tracking or geolocation data sharing without authorization.
  • 2.4.6 Financial information, social security numbers, or identity documents exposure.
  • 2.4.7 Medical records, health information, or genetic data unauthorized disclosure.
  • 2.4.8 Children's personal information collection or processing without parental consent.
  • 2.4.9 Workplace surveillance or employee monitoring without proper disclosure.
  • 2.4.10 Intimate or private images shared without subject's knowledge or consent.

2.5 Platform Abuse

  • 2.5.1 Spam, unsolicited bulk messages, or repetitive low-quality content.
  • 2.5.2 Attempts to circumvent content filters, rate limits, or safety measures.
  • 2.5.3 Creating multiple accounts to evade restrictions or manipulate systems.
  • 2.5.4 Automated abuse, bot networks, or coordinated inauthentic behavior.
  • 2.5.5 Reverse engineering, scraping, or unauthorized access to TNSA systems.
  • 2.5.6 Selling, transferring, or sharing account access without authorization.
  • 2.5.7 Using TNSA services to compete directly with TNSA or create substitute services.
  • 2.5.8 Excessive resource consumption or attempts to overload TNSA infrastructure.
  • 2.5.9 Manipulating or gaming recommendation algorithms or content ranking systems.
  • 2.5.10 Violating API terms, rate limits, or usage restrictions.

3. DETECTION & CLASSIFICATION SYSTEMS

3.1 Automated Detection

  • 3.1.1 Machine learning models trained on diverse datasets to identify policy violations.
  • 3.1.2 Natural language processing systems for text content analysis and classification.
  • 3.1.3 Computer vision algorithms for image and video content evaluation.
  • 3.1.4 Audio analysis systems for speech content and acoustic pattern detection.
  • 3.1.5 Multimodal AI systems for analyzing complex content across multiple formats.
  • 3.1.6 Real-time filtering systems integrated into all user-facing interfaces.
  • 3.1.7 Batch processing systems for comprehensive content review and analysis.
  • 3.1.8 Behavioral analysis systems for detecting patterns of abuse or manipulation.
  • 3.1.9 Network analysis tools for identifying coordinated inauthentic behavior.
  • 3.1.10 Continuous learning systems that adapt to new threats and violation patterns.

3.2 Human Review Process

  • 3.2.1 Trained content moderators review flagged content and edge cases.
  • 3.2.2 Subject matter experts evaluate specialized content requiring domain knowledge.
  • 3.2.3 Cultural and linguistic specialists review content in multiple languages and contexts.
  • 3.2.4 Senior reviewers handle appeals, complex cases, and policy interpretation questions.
  • 3.2.5 Quality assurance teams audit moderation decisions for consistency and accuracy.
  • 3.2.6 External expert consultations for highly specialized or sensitive content areas.
  • 3.2.7 Escalation procedures for content requiring legal or executive review.
  • 3.2.8 Regular training and calibration sessions for all human reviewers.
  • 3.2.9 Psychological support and wellness programs for content moderation staff.
  • 3.2.10 Continuous feedback loops between automated systems and human reviewers.

3.3 Classification Framework

  • 3.3.1 Severity levels: Critical, High, Medium, Low, and Informational classifications.
  • 3.3.2 Confidence scores indicating the certainty of automated detection systems.
  • 3.3.3 Context tags identifying the specific policy area or violation type.
  • 3.3.4 Risk assessments evaluating potential harm and appropriate response measures.
  • 3.3.5 Temporal classifications for time-sensitive content requiring immediate action.
  • 3.3.6 Geographic relevance tags for region-specific legal or cultural considerations.
  • 3.3.7 User impact assessments considering the affected audience and potential reach.
  • 3.3.8 Remediation recommendations specifying appropriate enforcement actions.
  • 3.3.9 Appeal eligibility determinations based on violation type and circumstances.
  • 3.3.10 Learning classifications to improve future detection and policy development.

This comprehensive Content Moderation Framework contains 200+ detailed policies and procedures. For specific enforcement and appeals information: