*1 United States Court of Appeals
FOR THE DISTRICT OF COLUMBIA CIRCUIT
Argued September 16, 2024 Decided December 6, 2024
No. 24-1113
T IK T OK I NC . AND B YTE D ANCE L TD .,
P ETITIONERS v.
M ERRICK B. G ARLAND , IN HIS OFFICIAL CAPACITY AS A TTORNEY G ENERAL OF THE U NITED S TATES , R ESPONDENT Consolidated with 24-1130, 24-1183
On Petitions for Review of Constitutionality of the Protecting Americans from Foreign Adversary Controlled Applications
Act Andrew J. Pincus argued the cause for TikTok Petitioners. With him on the briefs were Avi M. Kupfer , Alexander A. Berengaut , David M. Zionts , Megan A. Crowley , and John E. Hall .
Jeffrey L. Fisher argued the cause for Creator Petitioners. With him on the briefs were Ambika Kumar , Tim Cunningham , *2 Xiang Li , Elizabeth A. McNamara , Chelsea T. Kelly , James R. Sigel , Adam S. Sieff , and Joshua Revesz .
Jacob Huebert and Jeffrey M. Schwab were on the briefs for petitioner BASED Politics, Inc.
David Greene was on the brief for amici curiae Electronic Frontier Foundation, et al. in support of petitioners.
Jameel Jaffer and Eric Columbus were on the brief for amici curiae the Knight First Amendment Institute at Columbia University, et al. in support of petitioners.
Edward Andrew Paltzik and Serge Krimnus were on the brief for amicus curiae HungryPanda US, Inc. in support of petitioners.
Matt K. Nguyen , Travis LeBlanc , Robert H. Denniston , Kathleen R. Hartnett , and Jamie D. Robertson were on the brief for amici curiae Social and Racial Justice Community Nonprofits in support of petitioners.
Nicholas Reddick and Meryl Conant Governski were on the brief for amici curiae First Amendment Law Professors in support of petitioners.
Thomas A. Berry was on the brief for amicus curiae the Cato Institute in support of petitioners.
Mark Davies , Ethan L. Plail , and Edred Richardson were on the brief for amici curiae Professors Mueller, Edgar, Aaronson, and Klein in support of petitioners.
Aaron D. Van Oort was on the brief for amicus curiae Professor Matthew Steilen in support of petitioners.
Daniel Tenny , Attorney, U.S. Department of Justice, argued the cause for respondent. With him on the brief were Brian M. Boynton , Principal Deputy Assistant Attorney General, Brian D. Netter , Deputy Assistant Attorney General, Mark R. Freeman , Sharon Swingle , Casen B. Ross , Sean R. Janda , and Brian J. Springer , Attorneys, Matthew G. Olsen , Assistant Attorney General for National Security, Tyler J. Wood , Deputy Chief, Foreign Investment Review Section, and Tricia Wellman , Acting General Counsel, Office of the Director of National Intelligence.
Thomas R. McCarthy was on the brief for amici curiae Former National Security Officials in support of respondent.
Joel L. Thayer was on the brief for amici curiae Campaign for Uyghurs, et al. in support of respondent.
Joel L. Thayer was on the brief for amici curiae Zephyr Teachout, et al. in support of respondent.
Thomas M. Johnson, Jr. , Jeremy J. Broggi , and Joel S. Nolette were on the brief for amici curiae Chairman of the Select Committee on the CCP John R. Moolenaar, et al. in support of respondent.
David H. Thompson , Brian W. Barnes , and Megan M. Wold were on the brief for amicus curiae Professor D. Adam Candeub in support of respondent.
Thomas M. Johnson, Jr. , Jeremy J. Broggi , and Michael J. Showalter were on the brief for amici curiae Former Chairman of the Federal Communications Commission Ajit V. Pai and Former Assistant Secretary of the Treasury for Investment Security Thomas P. Feddo in support of respondent.
Jonathan Berry , Michael Buschbacher , Jared M. Kelson , James R. Conde , and William P. Barr were on the brief for amicus curiae American Free Enterprise Chamber of Commerce in support of respondent.
Austin Knudsen , Attorney General, Office of the Attorney General for the State of Montana, Christian B. Corrigan , Solicitor General, Peter M. Torstensen, Jr. , Deputy Solicitor General, Jason S. Miyares , Attorney General, Office of the Attorney General for the Commonwealth of Virginia, Erika L. Maley , Solicitor General, Kevin M. Gallagher , Principal Deputy Solicitor General, Steve Marshall , Attorney General, Office of the Attorney General for the State of Alabama, Treg Taylor , Attorney General, Office of the Attorney General for the State of Alaska, Tim Griffin , Attorney General, Office of the Attorney General for the State of Arkansas, Ashley Moody , Attorney General, Office of the Attorney General for the State of Florida, Christopher M. Carr , Attorney General, Office of the Attorney General for the State of Georgia, Raúl R. Labrador , Attorney General, Office of the Attorney General for the State of Idaho, Theodore E. Rokita , Attorney General, Office of the Attorney General for the State of Indiana, Brenna Bird , Attorney General, Office of the Attorney General for the State of Iowa, Russell Coleman , Attorney General, Office of the Attorney General for the Commonwealth of Kentucky, Liz Murrill , Attorney General, Office of the Attorney General for the State of Louisiana, Lynn Fitch , Attorney General, Office of the Attorney General for the State of Mississippi, Andrew Bailey , Attorney General, Office of the Attorney General for the State of Missouri, Michael T. Hilgers , Attorney General, Office of the Attorney General for the State of Nebraska, John M. Formella , Attorney General, Office of the Attorney General for the State of New Hampshire, Gentner F. Drummond , Attorney General, Office of the Attorney General for the State of Oklahoma, Alan Wilson , Attorney General, Office of the *5 Attorney General for the State of South Carolina, Marty J. Jackley , Attorney General, Office of the Attorney General for the State of South Dakota, Jonathan Skrmetti , Attorney General, Office of the Attorney General for the State of Tennessee, and Sean D. Reyes , Attorney General, Office of the Attorney General for the State of Utah, were on the brief for amici curiae State of Montana, Virginia, and 19 Other States in support of respondent.
Peter C. Choharis and Arnon D. Siegel were on the brief for amicus curiae the Foundation for Defense of Democracies in support of respondent.
Before: S RINIVASAN , Chief Judge , R AO , Circuit Judge , and G INSBURG , Senior Circuit Judge .
Opinion for the Court filed by Senior Circuit Judge G INSBURG .
Opinion concurring in part and concurring in the judgment filed by Chief Judge S RINIVASAN .
6
I. Background 8
A. The TikTok Platform 8 B. The Petitioners 9 C. National Security Concerns 11 D. The Act 15 1. Foreign adversary controlled applications 16 2. Prohibitions 18 3. The divestiture exemption 19 E. Procedural History 20 II. Analysis 20
A. Standing and Ripeness 21 B. The First Amendment 24 1. Heightened scrutiny applies. 24 2. The Act satisfies strict scrutiny. 32 a. The Government’s justifications are compelling. 33 (i) National security justifications 33 (ii) Data collection 38 (iii) Content manipulation 42 b. The Act is narrowly tailored. 48 (i) TikTok’s proposed NSA 49 (ii) Other options 53 (iii) Overinclusive / underinclusive 55 C. Equal Protection 57 D. The Bill of Attainder Clause 59 E. The Takings Clause 63 F. Alternative Relief 64 III. Conclusion
G INSBURG , Senior Circuit Judge : On April 24, 2024 the President signed the Protecting Americans from Foreign Adversary Controlled Applications Act into law. Pub. L. No. 118-50, div. H. The Act identifies the People’s Republic of China (PRC) and three other countries as foreign adversaries of the United States and prohibits the distribution or mainte- nance of “foreign adversary controlled applications.” [1] Its prohibitions will take effect on January 19, 2025 with respect to the TikTok platform.
Three petitions — filed by ByteDance Ltd. and TikTok, Inc.; Based Politics, Inc.; and a group of individuals (“Creators”) who use the TikTok platform — which we have consolidated, all present constitutional challenges to the Act. We conclude the portions of the Act the petitioners have stand- ing to challenge, that is the provisions concerning TikTok and its related entities, survive constitutional scrutiny. We therefore deny the petitions.
I. Background
This court has original and exclusive jurisdiction over this
case pursuant to Section 3 of the Act. The parties have submit-
ted several evidentiary appendices in support of their positions,
including sworn declarations from various experts. In review-
ing this material, we consider whether there is a genuine dis-
pute as to any material fact.
Cf.
Fed. R. Civ. P. 56(a), (c)(4).
Here, no dispute of “essential facts” stands in the way of our
deciding this case on the merits of the parties’ legal arguments.
See Cal. ex rel. State Lands Comm’n v. United States
, 457 U.S.
273, 278 (1982);
South Carolina v. Katzenbach
,
A. The TikTok Platform
TikTok is a social-media platform that lets users create, upload, and watch short video clips overlaid with text, voice- overs, and music. For each individual viewer, the platform creates a continuous sequence of videos based upon that user’s behavior and several other factors, with the aim of keeping that user engaged. The TikTok platform has approximately 170 million monthly users in the United States and more than one billion users worldwide.
What a TikTok user sees on the platform is determined by a recommendation engine, company content moderation deci- sions, and video promotion and filtering decisions. The recommendation engine is an algorithm that displays videos based upon content metadata and user behavior. It identifies a pool of candidate videos for a user, then scores and ranks those videos using machine-learning models designed to determine which video(s) would be most appealing to the user. The source code for the engine was originally developed by ByteDance, a company based in China that is the ultimate parent of TikTok. According to TikTok, the global TikTok team, which includes *9 Chinese engineers, “continually develop[s]” the recommenda- tion engine and platform source code. As we explain in more detail below, the recommendation engine for the version of the platform that operates in the United States is deployed to a cloud environment run by Oracle Corporation.
Content moderation decisions involve a combination of machine and human actions. According to TikTok every video on the TikTok platform goes through “automated moderation” and if deemed potentially problematic is sent to a human moderator for review. TikTok’s Head of Operations and Trust & Safety approves the “community guidelines” that drive content moderation on the platform.
Video promotion (also called “heating”) and demotion (also called “filtering”) decisions are used to advance TikTok’s commercial or other goals. These decisions involve promoting or limiting specific videos on the platform. According to TikTok, each video that is promoted is first reviewed by a human. Review teams are regionalized so that videos promoted in the United States are reviewed by U.S.-based reviewers. With respect to filtering, the platform follows “a set of rules to filter out and disperse certain content.” B. The Petitioners
Three groups of petitioners challenge the Act on constitu- tional grounds: ByteDance Ltd. and TikTok, Inc.; Based Politics, Inc.; and the self-styled Creators, eight individuals who use the TikTok platform. We refer to the latter two groups collectively as the User Petitioners. Where the corporate struc- ture of ByteDance affects our analysis, we identify the relevant corporate entity by name. Otherwise, we refer generally to the constellation of ByteDance entities as TikTok. Because PRC control of the TikTok platform is central to this case, we *10 provide the following overview of the relevant corporate relationships.
ByteDance Ltd., the ultimate parent company of TikTok, is incorporated in the Cayman Islands. The Government characterizes ByteDance as headquartered in China and ByteDance acknowledges that it has significant operations there. [2] ByteDance provides more than a dozen products through various operating subsidiaries, including Douyin, which is the counterpart to TikTok in China. The company was founded by Yiming Zhang, a Chinese national. Zhang retains 21 percent ownership of the company.
TikTok Ltd. is a wholly owned subsidiary of ByteDance and is also incorporated abroad. TikTok Ltd. operates the TikTok platform globally, except in China. The Government refers to TikTok entities that operate the platform outside the United States as “TikTok Global” and its U.S. operations as “TikTok US.”
TikTok Ltd. wholly owns TikTok LLC, which in turn wholly owns TikTok, Inc., a California corporation that pro- vides the TikTok platform to users in the United States. According to a TikTok declarant, TikTok’s “U.S. application and global application are highly integrated,” and the “global TikTok application itself is highly integrated with ByteDance.” Because the TikTok “platform and the content [are] global, the teams working on the platform, and the tools they use, neces- sarily must be, as well.” According to TikTok, one of ByteDance’s roles is “development of portions of the computer code that runs the TikTok platform.” In the Government’s view, TikTok “would try to comply if the PRC asked for specific actions to be taken to manipulate content for *11 censorship, propaganda, or other malign purposes on TikTok US.”
TikTok U.S. Data Security Inc. (TTUSDS) is a wholly owned subsidiary of TikTok, Inc., incorporated in Delaware. TikTok created TTUSDS to limit ByteDance’s access to the data of TikTok’s users in the United States and to monitor the security of the platform. TikTok represents that TTUSDS employees are separated from other TikTok employees, and that it partnered with Oracle to migrate the U.S. version of the TikTok platform into a cloud environment run by Oracle. TikTok also represents that TTUSDS and Oracle review updates to the platform made by ByteDance’s non-TTUSDS employees, and that Oracle has full access to TikTok’s source code. According to TikTok, TTUSDS is also responsible for deploying the recommendation engine in the United States, and TTUSDS signs off on any decision to promote or demote content in the United States.
C. National Security Concerns
As relevant here, the Executive [3] first became concerned about the PRC’s influence over TikTok in 2018 when ByteDance relaunched the platform in the United States following its acquisition of Musical.ly. In 2019, upon finding that “foreign adversaries” were “exploiting vulnerabilities in information and communications technology and services,” President Trump declared a national emergency. Securing the Information and Communications Technology and Services Supply Chain , Exec. Order No. 13873, 84 Fed. Reg. 22689, 22689 (May 15, 2019). Later that year, the Committee on Foreign Investment in the United States (CFIUS), which comprises the heads of several Executive Branch agencies, sent *12 a questionnaire to ByteDance about national security concerns related to ByteDance’s acquisition of Musical.ly. Thus began a lengthy investigatory process that culminated on August 1, 2020 with CFIUS concluding that TikTok could not suffi- ciently mitigate its national security concerns and referring the transaction to the President. The President, acting on that referral, ordered ByteDance to divest any “assets or property” that “enable or support ByteDance’s operation of the TikTok application in the United States.” Regarding the Acquisition of Musical.ly by ByteDance Ltd. , 85 Fed. Reg. 51297, 51297 (Aug. 14, 2020).
President Trump separately invoked his powers under the
International Emergency Economic Powers Act (IEEPA) and
the National Emergencies Act to address “the threat posed by
one mobile application in particular, TikTok.”
Addressing the
Threat Posed by TikTok
, Exec. Order No. 13942, 85 Fed. Reg.
48637, 48637 (Aug. 6, 2020). President Trump prohibited
certain “transactions” with ByteDance or its subsidiaries,
id.
at
48638, and the Secretary of Commerce later published a list of
prohibited transactions, 85 Fed. Reg. 60061 (Sept. 24, 2020).
Litigation ensued, and two courts enjoined the President’s
prohibitions under the IEEPA as exceeding his authority under
that law.
TikTok Inc. v. Trump
, 507 F. Supp. 3d 92, 102
(D.D.C. 2020);
Maryland v. Trump
,
In 2021, President Biden withdrew President Trump’s IEEPA executive order and issued a new one. In the new order, the President identified the PRC as “a foreign adversary” that “continues to threaten the national security, foreign policy, and economy of the United States” through its control of “software applications” used in the United States. Protecting Americans’ Sensitive Data From Foreign Adversaries , Exec. Order No. 14034, 86 Fed. Reg. 31423, 31423 (June 9, 2021). President *13 Biden elaborated that “software applications” can provide foreign adversaries with “vast swaths of information from users,” and that the PRC’s “access to large repositories” of such data “presents a significant risk.” Id. President Biden directed several executive agencies to provide risk mitigation options, and he asked for recommended “executive and legislative actions” to counter risks “associated with connected software applications that are designed, developed, manufactured, or supplied by persons owned or controlled by, or subject to the jurisdiction or direction of, a foreign adversary.” Id. The following year, President Biden signed into law a bill prohibit- ing the use of TikTok on government devices. See generally Pub. L. No. 117-328, div. R, 136 Stat. 5258 (2022).
Litigation regarding President Trump’s divestiture order pursuant to CFIUS’s referral began when TikTok filed suit in this court challenging the constitutionality of the order. See Pet. for Review, TikTok Inc. v. CFIUS , No. 20-1444 (2020). At the request of the parties, in February 2021 this court placed that case in abeyance while the new administration considered the matter and the parties negotiated over an alternative remedy that would sufficiently address the Executive’s national secu- rity concerns.
During 2021 and 2022, TikTok submitted multiple drafts of its proposed National Security Agreement (NSA) and Executive Branch officials held numerous meetings to consider TikTok’s submissions. According to TikTok, there were “at least” 14 meetings or calls, nine written presentations by TikTok, and 15 email exchanges in which “CFIUS posed ques- tions related to [TikTok’s] operations and the NSA terms.” A TikTok declarant describes the negotiations as “protracted, detailed, and productive,” and the Government similarly characterizes them as “significant” and “intensive.” Also as part of the process, “Executive Branch negotiators engaged in *14 extensive, in-depth discussions with Oracle, the proposed Trusted Technology Provider, whose responsibility under the proposed mitigation structure included storing data in the United States, performing source code review, and ensuring safety of the operation of the TikTok platform in the United States.”
In August 2022, TikTok submitted its last proposal. Although the parties dispute certain details about how to inter- pret specific provisions, the broad contours of TikTok’s pro- posed NSA are undisputed. Three aspects of the proposal bear emphasis.
First, the proposal purported to give TikTok operational independence from ByteDance by creating a new entity insu- lated from the influence of ByteDance, namely TTUSDS. The key management personnel of TTUSDS were to be subject to approval by the Government.
Second, the proposed NSA would create three tiers of data to limit the ability of ByteDance to access the data of TikTok’s users in the United States. Protected Data generally would encompass personal information about TikTok’s U.S. users — such as their usernames, passwords, user-created content, and any other personally identifiable information — unless such data were classified as Excepted Data or Public Data. Sharing of Protected Data with ByteDance would be prohibited except pursuant to limited-access protocols. Excepted Data would include data that platform users authorized to be shared with TikTok or its affiliates; certain defined data fields; and encrypted usernames, phone numbers, email addresses, etc., for routing to the United States. Public Data would include data generally accessible to platform users, as well as any content a user decides to make public. Under the proposed NSA, TikTok could send Excepted Data and Public Data to ByteDance.
Third, the proposal provided for a “trusted third party,” Oracle, to inspect the source code, including TikTok’s recommendation engine. It also gave the Government author- ity, under certain circumstances, to instruct TikTok to shut down the platform in the United States, which TikTok calls a “kill switch.”
The Executive determined the proposed NSA was insuffi- cient for several reasons. Most fundamentally, certain data of U.S. users would still flow to China and ByteDance would still be able to exert control over TikTok’s operations in the United States. The Executive also did not trust that ByteDance and TTUSDS would comply in good faith with the NSA. Nor did the Executive have “sufficient visibility [into] and resources to monitor” compliance. In the Executive’s view, divestment was the only solution that would adequately address its national security concerns. TikTok nevertheless voluntarily imple- mented some of its proposed mitigation measures.
D. The Act
In the months leading to passage of the Act, the Congress conducted a series of classified briefings and hearings regard- ing the Government’s national security concerns. The Congress then debated and passed the Act as one part of a broader appropriations bill, which also included the Protecting Americans’ Data from Foreign Adversaries Act of 2024, Pub. L. No. 118-50, div. I (2024), hereinafter the Data Broker Law. The Act and the Data Broker Law include nearly identical definitions of “foreign adversary country” and “controlled by a foreign adversary.” Their aims also overlap. Section 2(a) of the Data Broker Law prohibits third party data brokers from transferring “personally identifiable sensitive data of a United States individual” to a foreign adversary country or an entity “controlled by a foreign adversary.” The Act complements that *16 provision by limiting the ability of foreign adversaries to collect data directly through adversary controlled applications.
The Act itself is narrowly constructed to counter foreign adversary control through divestiture. Three aspects of the Act are particularly relevant to this case: (1) the definition of for- eign adversary controlled applications, (2) prohibitions in the Act, and (3) the divestiture option.
1. Foreign adversary controlled applications The Act defines a Foreign Adversary Controlled Application as “a website, desktop application, mobile applica- tion, or augmented or immersive technology application that is operated, directly or indirectly” by either of two distinct groups. § 2(g)(3). The first group consists of the ByteDance constellation of entities, including TikTok, which is identified by name. § 2(g)(3)(A). The second group consists of every cov- ered company [4] that is determined by the President to present a *17 significant threat to national security. Specifically, it includes any “covered company” that:
(i) is controlled by a foreign adversary; [5] and (ii) that is determined by the President to present a significant threat to the national security of the United States following the issuance of — (I) a public notice proposing such determination; and (II) a public report to Congress, submitted not less than 30 days before such determination, describing the specific national security concern involved and containing a classified annex and a description of application, or augmented or immersive technology
application. § 2(g)(2)(A). The term excludes, however, entities that operate an “application whose primary purpose is to allow users to post product reviews, business reviews, or travel information and reviews.” § 2(g)(2)(B).
[5] The term “controlled by a foreign adversary” means a “covered company or other entity” that is:
(A) a foreign person that is domiciled in, is headquartered in, has its principal place of business in, or is organized under the laws of a foreign adversary country; (B) an entity with respect to which a foreign person or combination of foreign persons described in subpara- graph (A) directly or indirectly own at least a 20 percent stake; or
(C) a person subject to the direction or control of a foreign person or entity described in subparagraph (A) or (B). § 2(g)(1). The definition of “foreign adversary country” encom- passes China, Russia, Iran, and North Korea. § 2(g)(2) (defining the term by reference to 10 U.S.C. § 4872(d)(2)).
what assets would need to be divested to execute a qualified divestiture.
§ 2(g)(3)(B).
2. Prohibitions
The Act contains prohibitions, § 2(a), and a “data and information portability” requirement, § 2(b). The prohibitions do not directly proscribe conduct by an entity that owns a for- eign adversary controlled application. Instead, they bar others from providing critical support in the United States for such an application. Specifically, the Act makes it “unlawful for an entity to distribute, maintain, or update” a foreign adversary controlled application in any of two ways:
(A) Providing services to distribute, maintain, or update such foreign adversary controlled applica- tion (including any source code of such applica- tion) by means of a marketplace (including an online mobile application store) through which users within the land or maritime borders of the United States may access, maintain, or update such application.
(B) Providing internet hosting services to enable the distribution, maintenance, or updating of such for- eign adversary controlled application for users within the land or maritime borders of the United States.
§ 2(a)(1).
With respect to TikTok, the prohibitions take effect 270 days after the Act was passed into law, that is, on January 19, 2025. § 2(a)(2)(A). With respect to applications subject to the generally applicable provisions, the prohibitions take effect 270 days after “the relevant determination of the President.” *19 § 2(a)(2)(B). In both situations, the President can grant a one- time, 90-day extension under specific circumstances not rele- vant here. § 2(a)(3).
Failure to comply with the Act can result in substantial monetary penalties. § 2(d)(1). To enforce the Act the Attorney General, following an investigation, can file suit in an appropri- ate district court. § 2(d)(2).
3. The divestiture exemption Section 2(c) of the Act provides an exemption “for quali- fied divestitures.” That is, the prohibitions do not apply if “a qualified divestiture is executed before the date on which a prohibition under subsection (a) would begin to apply.” § 2(c)(1)(A). If a qualified divestiture is executed after that date, then the prohibitions “shall cease to apply.” § 2(c)(1)(B). A “qualified divestiture” is defined as a transaction that:
(A) the President determines, through an interagency process, would result in the relevant foreign adversary controlled application no longer being controlled by a foreign adversary; and (B) the President determines, through an interagency process, precludes the establishment or mainte- nance of any operational relationship between the United States operations of the relevant foreign adversary controlled application and any formerly affiliated entities that are controlled by a foreign adversary, including any cooperation with respect to the operation of a content recommendation algo- rithm or an agreement with respect to data sharing. § 2(g)(6).
E. Procedural History
This case concerns three petitions challenging the Act that this court consolidated for review. On May 17, 2024 the parties jointly asked this Court to expedite the case. The parties advised that they intended to append evidentiary materials to their briefs. The Government noted that it was evaluating the need to file an ex parte evidentiary submission given the classified material implicated by the case. The petitioners reserved the right to object to any such submission.
The parties ultimately submitted evidence with their briefs. TikTok’s submission included several expert declara- tions as well as a declaration from its Head of Operations and Trust & Safety. The User Petitioners filed declarations underscoring the diverse ways in which they use the TikTok platform. The Government filed declarations explaining its national security concerns and why it found TikTok’s proposed NSA insufficient to meet those concerns. TikTok filed rebuttal declarations with its reply brief.
Portions of the Government’s brief and evidentiary submission were redacted because they contain classified information. The Government filed a motion requesting leave to file unredacted versions of its brief and supporting evidence under seal and ex parte, which documents the Government later lodged with this court. The petitioners opposed the Government’s motion and alternatively moved this court to appoint a special master and issue a temporary injunction in order to mitigate prejudice arising from the Government’s classified filings.
II. Analysis
The petitioners seek a declaratory judgment that the Act violates the Constitution and an order enjoining the Attorney *21 General from enforcing it. Because the petitioners are bringing a pre-enforcement challenge to the Act, we must determine the extent to which this court can consider their claims consistent with the standing aspect of the “case or controversy” require- ment of Article III of the Constitution. We conclude that TikTok has standing to challenge those portions of the Act that directly affect the activities of ByteDance and its affiliates. We further conclude that TikTok’s challenge to those portions of the Act is ripe.
On the merits, we reject each of the petitioners’ constitu- tional claims. As we shall explain, the parts of the Act that are properly before this court do not contravene the First Amendment to the Constitution of the United States, nor do they violate the Fifth Amendment guarantee of equal protection of the laws; constitute an unlawful bill of attainder, in violation of Article I, § 9, clause 3; or work an uncompensated taking of private property in violation of the Fifth Amendment. A. Standing and Ripeness
We have an independent duty to assure ourselves that the
petitioners and their claims satisfy the requirements of Article
III.
Exelon Corp. v. FERC
,
“To establish standing for a pre-enforcement challenge, a
plaintiff must demonstrate first an intention to engage in a
course of conduct arguably affected with a constitutional inter-
est, but proscribed by a statute and, second, that there exists a
credible threat of prosecution thereunder.”
Muthana v.
Pompeo
,
Ripeness is “related” but focuses “on the timing of the
action rather than on the parties seeking to bring it.”
Navegar,
Inc. v. United States
,
TikTok and its claims challenging enforcement of the prohibitions of the Act based upon the TikTok-specific provi- sions clearly satisfy the requirements respectively for standing and ripeness. The prohibitions based upon those provisions *23 take effect by operation of law on January 19, 2025. After that date, third parties that make the TikTok platform available in the United States would run a significant risk of incurring monetary penalties under § 2(d)(1). Even if the Act went unen- forced, the risk of penalties alone could cause third parties to suspend support for the TikTok platform, such as by removing it from online marketplaces, and an injunction would prevent that harm. TikTok therefore has Article III standing to pursue its claims.
The ripeness inquiry is likewise straightforward. TikTok risks severe hardship from delayed review, and we have an ade- quate record on which to resolve the company’s challenges to the constitutionality of the TikTok-specific provisions of the Act.
To the extent TikTok seeks to enjoin future enforcement of the prohibitions under the generally applicable track, TikTok does not have standing. Nor if it did would such a request be ripe for judicial review. Recall that applying the prohibitions under the generally applicable framework requires certain procedural steps and a presidential determination pursuant to § 2(g)(3)(B). Those steps include public notice, a description of the national security concern, a classified annex, and a description of assets to be divested. § 2(g)(3)(B)(ii). The President has not invoked those procedures with respect to TikTok (or any other company), and it would be self-evidently premature for the court even to consider a request for an injunc- tion against the President ever doing so. We consequently limit our analysis to the constitutionality of the Act as applied to the TikTok-specific provisions that will go into effect next month. [7] *24 B. The First Amendment
This case requires that we apply longstanding First
Amendment principles to somewhat novel facts: A popular
social-media platform, subject to the control of a foreign adver-
sary nation, that a statute requires be divested because of
national security risks. The issue is made more complex by the
web of subsidiaries wholly owned by ByteDance that lie
behind the TikTok platform.
See Moody v. NetChoice, LLC
,
We conclude the Act implicates the First Amendment and is subject to heightened scrutiny. Whether strict or intermediate scrutiny applies is a closer question. The relevant portions of the Act are facially content neutral, but the Government argua- bly based its content-manipulation justification for the Act upon the content on the platform. We think it only prudent, therefore, to assume without deciding that the higher standard applies.
1. Heightened scrutiny applies.
As in most First Amendment cases, the parties spend much of their time debating the appropriate standard of review. The petitioners urge the court to apply strict scrutiny but contend the Act fails intermediate scrutiny as well. The Government suggests we apply only rational basis review, alternatively advocates intermediate scrutiny, but maintains the Act satisfies even strict scrutiny.
for at least one plaintiff, we need not consider the standing of the other plaintiffs to raise that claim” (cleaned up)).
Under intermediate scrutiny, the Act complies with the
First Amendment “if it advances important governmental inter-
ests unrelated to the suppression of free speech and does not
burden substantially more speech than necessary to further
those interests.”
Turner Broad. Sys., Inc. v. FCC
(
Turner II
),
We think it clear that some level of heightened scrutiny is required. The question whether intermediate or strict scrutiny applies is difficult because the TikTok-specific provisions are facially content neutral, yet the Government justifies the Act in substantial part by reference to a foreign adversary’s ability to manipulate content seen by Americans. No Supreme Court case directly addresses whether such a justification renders a law content based, thereby triggering strict scrutiny. There are reasonable bases to conclude that intermediate scrutiny is appropriate even under these circumstances. We need not, however, definitively decide that question because we con- clude the Act “passes muster even under the more demanding standard.” FEC v. Int’l Funding Inst. , 969 F.2d 1110, 1116 (D.C. Cir. 1992); see also In re Sealed Case , 77 F.4th 815, 829–30 (D.C. Cir. 2023) (assuming without deciding that strict scrutiny applied).
At the outset, we reject the Government’s ambitious argu-
ment that this case is akin to
Arcara v. Cloud Books, Inc.
,
478 U.S. 697 (1986), and does not implicate the First
Amendment at all. That case concerned enforcement of “a
public health regulation of general application against” an adult
bookstore being “used for prostitution.”
Id.
at 707.
*26
Enforcement of a generally applicable law unrelated to
expressive activity does not call for any First Amendment
scrutiny.
Id.
By contrast, the First Amendment is implicated in
“cases involving governmental regulation of conduct that has
an expressive element,” or when a statute is directed at an
activity without an expressive component but imposes “a
disproportionate burden upon those engaged in protected First
Amendment activities.”
Id.
at 703–04;
see also Alexander v.
United States
,
Here the Act imposes a disproportionate burden on
TikTok, an entity engaged in expressive activity. The
Government concedes, as it must after
NetChoice
, that the
curation of content on TikTok is a form of speech. 144 S. Ct. at
2401. Like the social media companies in that case, TikTok
delivers a “personalized collection” of content to users and
moderates this content pursuant to its community guidelines.
Id.
at 2403–04. The Act plainly “single[s] out” that expressive
activity by indirectly subjecting TikTok — and so far, only
TikTok — to the divestiture requirement.
Arcara
, 478 U.S. at
707;
cf. Nat’l Rifle Ass’n of Am. v. Vullo
,
The Government suggests that because TikTok is wholly owned by ByteDance, a foreign company, it has no First Amendment rights. Cf. Agency for Int’l Dev. v. All. for Open Soc’y Int’l, Inc. , 591 U.S. 430, 436 (2020) (explaining that “foreign organizations operating abroad have no First Amendment rights”). TikTok, Inc., however, is a domestic entity operating domestically. See NetChoice , 144 S. Ct. at 2410 (Barrett, J., concurring) (identifying potential “complexi- ties” for First Amendment analysis posed by the “corporate structure and ownership of some platforms”). The Government does not dispute facts suggesting at least some of the regulated speech involves TikTok’s U.S. entities. See TikTok App. 811– 12, 817–18 (explaining that promoted videos are “reviewed by a U.S.-based reviewer,” that an executive employed by a U.S. entity approves the guidelines for content moderation, and that the recommendation engine “is customized for TikTok’s vari- ous global markets” and “subject to special vetting in the United States”).
Nor does the Government argue we should “pierce the
corporate veil” or “invoke any other relevant exception” to the
fundamental principle of corporate separateness.
Agency for
Int’l Dev.
,
The next question is whether intermediate or strict scrutiny
is appropriate, which turns on whether the Act is content
*28
neutral or content based.
See Turner Broad. Sys., Inc. v. FCC
(
Turner I
),
The provisions of the Act before us are facially content
neutral because they do not target speech based upon its
communicative content. The TikTok-specific provisions
instead straightforwardly require only that TikTok divest its
platform as a precondition to operating in the United States. On
its face, the Act concerns control by a foreign adversary and
not “the topic discussed or the idea or message expressed.”
City
of Austin
,
TikTok insists the TikTok-specific provisions nonetheless
require strict scrutiny because they single out a particular
speaker. To be sure, laws that “discriminate among media, or
among different speakers within a single medium, often present
serious First Amendment concerns.”
Turner I
,
Whether the Act, which is facially content neutral, is
subject to strict scrutiny therefore turns upon the Government’s
justifications for the law.
See Ward v. Rock Against Racism
,
At the same time, the Government’s concern with content
manipulation does not reflect “an impermissible purpose or
justification.”
City of Austin
,
Although we can conceive of reasons intermediate scru-
tiny may be appropriate under these circumstances, we ulti-
mately do not rest our judgment on those reasons because the
Act satisfies “the more demanding standard.”
Int’l Funding
Inst.
,
2. The Act satisfies strict scrutiny.
To satisfy strict scrutiny the Government must “demon-
strate that a speech restriction: (1) serves a compelling
government interest; and (2) is narrowly tailored to further that
interest.”
In re Sealed Case
,
We emphasize from the outset that our conclusion here is fact-bound. The multi-year efforts of both political branches to investigate the national security risks posed by the TikTok platform, and to consider potential remedies proposed by TikTok, weigh heavily in favor of the Act. The Government has offered persuasive evidence demonstrating that the Act is narrowly tailored to protect national security. “Given the sensi- tive interests in national security and foreign affairs at stake,” the Government’s judgment based upon this evidence “is enti- tled to significant weight.” Holder v. Humanitarian Law Project , 561 U.S. 1, 36 (2010). Our deference to the Government’s national-security assessment “is redoubled by the repeated acts of” the political branches to address the national security problems presented by the TikTok platform. Hikvision USA, Inc. v. FCC , 97 F.4th 938, 948 (D.C. Cir. 2024). The Act was the culmination of extensive, bipartisan action by the Congress and by successive presidents. It was carefully crafted to deal only with control by a foreign adver- sary, and it was part of a broader effort to counter a well- substantiated national security threat posed by the PRC. Under *33 these circumstances, the provisions of the Act that are before us withstand the most searching review.
a. The Government’s justifications are compelling.
Recall that the Government offers two national security justifications for the Act: to counter (1) the PRC’s efforts to collect data of and about persons in the United States, and (2) the risk of the PRC covertly manipulating content on TikTok. Each constitutes an independently compelling national security interest.
In reaching that conclusion, we follow the Supreme Court
in affording great weight to the Government’s “evaluation of
the facts” because the Act “implicates sensitive and weighty
interests of national security and
foreign affairs.”
Humanitarian Law Project
, 561 U.S. at 33–34;
Trump v.
Hawaii
, 585 U.S. 667, 707–08 (2018) (same);
see, e.g.
,
Pac.
Networks Corp. v. FCC
,
(i)
National security justifications
The Government provides persuasive support for its
concerns regarding the threat posed by the PRC in general and
*34
through the TikTok platform in particular. As Assistant
Director of National Intelligence Casey Blackburn explained,
the “PRC is the most active and persistent cyber espionage
threat to U.S. government, private-sector, and critical
infrastructure networks.” Its hacking program “spans the
globe” and “is larger than that of every other major nation,
combined.” The PRC has “pre-positioned” itself “for potential
cyber-attacks against U.S. critical infrastructure by building
out offensive weapons within that infrastructure.” Consistent
with that assessment, the Government “has found persistent
PRC access in U.S. critical telecommunications, energy, water,
and other infrastructure.”
See China Telecom (Ams.) Corp. v.
FCC
,
Of particular relevance to the Government’s first justifica- tion for the Act, the PRC has engaged in “extensive and years- long efforts to accumulate structured datasets, in particular on U.S. persons, to support its intelligence and counterintelligence operations.” It has done so through hacking operations, such as by penetrating the U.S. Government Office of Personnel Management’s systems and taking “reams” of personal data, stealing financial data on 147 million Americans from a credit- reporting agency, and “almost certainly” extracting health data on nearly 80 million Americans from a health insurance provider.
The PRC’s methods for collecting data include using “its relationships with Chinese companies,” making “strategic investments in foreign companies,” and “purchasing large data sets.” For example, the PRC has attempted “to acquire sensitive *35 health and genomic data on U.S. persons” by investing in firms that have or have access to such data. Government counterintelligence experts describe this kind of activity as a “hybrid commercial threat.”
The PRC poses a particularly significant hybrid commer-
cial threat because it has adopted laws that enable it to access
and use data held by Chinese companies.
See China Telecom
(Ams.) Corp.
,
The upshot of these and other laws, according to the Government’s declarants, is that “even putatively ‘private’ companies based in China do not operate with independence from the government and cannot be analogized to private companies in the United States.” Through its “control over Chinese parent companies,” the PRC can also “access infor- mation from and about U.S. subsidiaries and compel their cooperation with PRC directives.” As a result, the PRC can “conduct espionage, technology transfer, data collection, and other disruptive activities under the disguise of an otherwise legitimate commercial activity.” According to Kevin Vorndran, Assistant Director of the FBI’s Counterintelligence Division, the PRC endeavors strategically to pre-position commercial entities in the United States that the PRC can later “co-opt.” These pre-positioning “tactics can occur over the span of several years of planning and implementation, and they *36 are one “part of the PRC’s broader geopolitical and long-term strategy to undermine U.S. national security.”
The PRC likewise uses its cyber capabilities to support its influence campaigns around the world. Those global “influence operations” aim to “undermine democracy” and “extend the PRC’s influence abroad.” Specifically, the PRC conducts “cyber intrusions targeted to affect U.S. and non-U.S. citizens beyond its borders — including journalists, dissidents, and individuals it views as threats — to counter and suppress views it considers critical of [the PRC].” Notably, the Government reports that “ByteDance and TikTok Global have taken action in response to PRC demands to censor content outside of China.”
As it relates to TikTok in the United States, the Government predicts that ByteDance and TikTok entities “would try to comply if the PRC asked for specific actions to be taken to manipulate content for censorship, propaganda, or other malign purposes on TikTok US.” The Government says that ByteDance, which is subject to PRC laws requiring cooperation with the PRC, could do so by acting unilaterally or by conscripting its U.S. entities. The former conclusion is evi- denced by the fact that the PRC maintains a powerful Chinese Communist Party committee “embedded in ByteDance” through which it can “exert its will on the company.” As of 2022, that committee “was headed by the company’s chief editor and comprised at least 138 employees at its Beijing office, including senior company managers.” The latter conclu- sion is supported by the fact that TikTok’s U.S. operations are “heavily reliant” on ByteDance. As TikTok’s declarants have put it, “TikTok in the United States is an integrated part of the global platform” supported by teams “spread across several different corporate entities and countries,” and TikTok is “highly integrated with ByteDance.”
The Government also identifies several public reports, which were considered by the Congress prior to passing the Act, regarding the risks posed by TikTok. [9] For example, a Government declarant points to “reporting by Forbes Magazine” to illustrate in part why the Government did not trust TikTok’s proposed mitigation measures. The reporting suggested “that ByteDance employees abused U.S. user data, even after the establishment of TTUSDS,” and drew attention to “audio recordings of ByteDance meetings” that indicated “ByteDance retained considerable control and influence over TTUSDS operations.” In its report recommending passage of the Act, a committee of the Congress collected “a list of public statements that have been made regarding the national security risks posed by . . . TikTok.” H.R. Rep. No. 118-417, at 5–12 (2024). According to the committee, public reporting sug- gested that TikTok had stored sensitive information about U.S. persons (including “Social Security numbers and tax identifica- tions”) on servers in China; TikTok’s “China-based employ- ees” had “repeatedly accessed non-public data about U.S. TikTok users”; ByteDance employees had “accessed TikTok user data and IP addresses to monitor the physical locations of *38 specific U.S. citizens”; and PRC agents had inspected “TikTok’s internal platform.” Id. at 7–10.
The resulting judgment of the Congress and the Executive regarding the national security threat posed by the TikTok platform “is entitled to significant weight, and we have persua- sive evidence [in the public record] before us to sustain it.” Humanitarian Law Project , 561 U.S. at 36. The petitioners raise several objections to each national security justification, which we take up next, but the bottom line is that they fail to overcome the Government’s considered judgment and the deference we owe that judgment.
(ii)
Data collection
TikTok disputes certain details about the Government’s
concern with its collection of data on U.S. persons but misses
the forest for the trees. The TikTok platform has more than 170
million monthly users in the United States. It is an immensely
popular platform on which users in the United States have
uploaded more than 5.5 billion videos in a single year.
According to TikTok’s “privacy policy,” TikTok automatically
collects large swaths of data about its users, including device
information (IP address, keystroke patterns, activity across
devices, browsing and search history, etc.) and location data
(triangulating SIM card or IP address data for newer versions
of TikTok and GPS information for older versions). TikTok,
Privacy Policy
, https://perma.cc/E36Q-M3KS (last updated
Aug. 19, 2024). It may also collect image and audio infor-
mation (including biometric identifiers and biometric infor-
mation such as faceprints and voiceprints); metadata (describ-
ing how, when, where, and by whom content was created, col-
lected, or modified); and usage information (including content
that users upload to TikTok).
Id.
That is not to mention infor-
mation that users voluntarily provide, such as name, age,
*39
username, password, email, phone number, social media
account information, messages exchanged on the platform and,
“with your permission,” your “phone and social network con-
tacts.”
Id.
TikTok’s “privacy policy” also makes clear that it
uses these data to “infer additional information” about its users.
Given the magnitude of the data gathered by TikTok and
TikTok’s connections to the PRC, two consecutive presidents
understandably identified TikTok as a significant vulnerability.
Access to such information could, for example, allow the PRC
to “track the locations of Federal employees and contractors,
build dossiers of personal information for blackmail, and
conduct corporate espionage.”
Addressing the Threat Posed by
TikTok
,
TikTok does not deny that it collects a substantial amount of data on its users. Instead TikTok disputes details about the Government’s understanding of its data practices and questions the sincerity of the Government’s data justification. At the same time, however, TikTok’s own declarants provide support for the Government’s concern. They emphasize the integrated nature of the TikTok platform to argue that divestiture would be infeasible. They argue that prohibiting data sharing between TikTok in the United States and “the entities that operate the global platform” would make TikTok uncompetitive with “rival, global platforms.” They also acknowledge that, even under TikTok’s proposed NSA, ByteDance would continue to have access to some Protected Data on TikTok users in the United States through “limited access protocols.” They like- wise state that TikTok’s proposed NSA “does allow for TTUSDS and Oracle to send ‘Excepted Data’ to ByteDance.”
Set against those statements, TikTok’s arguments concerning the specific data collected and TikTok’s voluntary data protection efforts fall flat. For example, TikTok quibbles with the Government’s stated concern that TikTok collects data *40 on users’ “precise locations, viewing habits, and private mes- sages,” including “data on users’ phone contacts who do not themselves use TikTok.” Gov’t Br. 1; see TikTok Reply Br. 25. According to a TikTok declarant, the current version of TikTok can only “approximate users’ geographic locations.” Access to a user’s contact list, likewise, is currently available only if a user affirmatively opts in, and it is “anonymized and used only to facilitate connections with other TikTok users.” TikTok Reply Br. 25. TikTok further points to other data protections that it claims to provide, such as storing sensitive user data in the United States and controlling access to them.
The Government’s data-related justification for the Act, however, does not turn on the details of TikTok’s mitigation measures. Even after extended negotiations, TikTok could not satisfactorily resolve the Government’s concerns. We have no doubt, and the Government has never denied, that TikTok’s proposed NSA would mitigate the Government’s concerns to some extent. Nor do we doubt that TikTok’s voluntary mitiga- tion efforts provide some protection. The problem for TikTok is that the Government exercised its considered judgment and concluded that mitigation efforts short of divestiture were insufficient, as a TikTok declarant puts it, to mitigate “risks to acceptable levels.” At bottom, the Government lacks confi- dence that it has sufficient visibility and resources to monitor TikTok’s promised measures, nor does it have “the requisite trust” that “ByteDance and TTUSDS would comply in good faith.” The court can neither fault nor second guess the Government on these crucial points.
This situation is much like that in Pacific Networks Corp. v. FCC , 77 F.4th 1160 (D.C. Cir. 2023), which involved the Executive’s decision to revoke authorizations held by PRC- controlled companies to operate communication lines in the United States. There, as here, the PRC indirectly controlled the *41 companies “through a web of foreign affiliates.” 77 F.4th at 1163 (cleaned up). The Executive “concluded that China’s ownership raised significant concerns that the [companies] would be forced to comply with Chinese government requests.” Id. (cleaned up). The Government was concerned that the PRC could “access, monitor, store, and in some cases disrupt or misroute U.S. communications, which in turn [would] allow them to engage in espionage and other harmful activities against the United States.” Id. (cleaned up). The Executive “further concluded that the [companies] had shown a lack of candor and trustworthiness” and therefore “nothing short of revocation would ameliorate the national-security risks.” Id. This court declined the appellants’ invitation to “second-guess” the Executive’s judgment regarding the threat to national security. Id. at 1164. We also upheld the Executive’s conclusion that the companies’ “untrustworthiness would make any mitigation agreement too risky” in part because the Executive could not “comprehensively monitor compliance” or “reliably detect surreptitious, state-sponsored efforts at evasion.” Id. at 1165–66. The same considerations similarly support the Government’s judgment here.
We also reject TikTok’s argument that the Government’s
data-related concerns are speculative. The Government “need
not wait for a risk to materialize” before acting; its national
security decisions often must be “based on informed judg-
ment.”
China Telecom (Ams.) Corp.
,
TikTok next contends that, because other companies with operations in China collect data in the United States, its data collection is not the Government’s real concern. As already explained, however, the Act complements the Data Broker Law, which limits the access of any foreign adversary country (or entity controlled by such a country) to data from third-party brokers. The Act also includes a generally applicable frame- work through which the Executive can address other foreign adversary controlled applications in the future. That the Act does not fully solve the data collection threat posed by the PRC does not mean it was not a step in the right direction. Moreover, TikTok does not identify any company operating a comparable platform in the United States with equivalent connections to the PRC. Nor would it be dispositive if TikTok had done so because the political branches are free to “focus on their most pressing concerns.” Williams-Yulee v. Florida Bar , 575 U.S. 433, 449 (2015). The Government’s multi-year efforts to address the risks posed by the TikTok platform support the conclusion that TikTok was, in fact, the Government’s most pressing concern.
(iii) Content manipulation Preventing covert content manipulation by an adversary nation also serves a compelling governmental interest. The petitioners object for two reasons, neither of which persuades.
First, TikTok incorrectly frames the Government’s
justification as suppressing propaganda and misinformation.
The Government’s justification in fact concerns the risk of the
PRC covertly manipulating content on the platform. For that
reason, again, the Act is directed only at control of TikTok by
*43
a foreign adversary nation. At points, TikTok also suggests the
Government does not have a legitimate interest in countering
covert content manipulation by the PRC. To the extent that is
TikTok’s argument, it is profoundly mistaken. “At the heart of
the First Amendment lies the principle that each person should
decide for himself or herself the ideas and beliefs deserving of
expression, consideration, and adherence. Our political system
and cultural life rest upon this ideal.”
Turner I
,
In this case, a foreign government threatens to distort free
speech on an important medium of communication. Using its
hybrid commercial strategy, the PRC has positioned itself to
manipulate public discourse on TikTok in order to serve its
own ends. The PRC’s ability to do so is at odds with free speech
fundamentals. Indeed, the First Amendment precludes a
domestic government from exercising comparable control over
a social media company in the United States.
See NetChoice
,
Like the Supreme Court, “We also find it significant that
[the Government] has been conscious of its own responsibility
to consider how its actions may implicate constitutional
*44
concerns.”
Humanitarian Law Project
,
Consequently, the Act is not, as the User Petitioners sug-
gest, an effort to “control the flow of ideas to the public.”
Lamont v. Postmaster Gen.
,
TikTok resists this conclusion by emphasizing stray com-
ments from the congressional proceedings that suggest some
congresspersons were motivated by hostility to certain content.
The Supreme Court, however, has repeatedly instructed that
courts should “not strike down an otherwise constitutional
statute on the basis of an alleged illicit legislative motive.”
O’Brien
,
The petitioners nevertheless contend the divestiture provi- sions and an exclusion from the generally applicable track betray the Government’s real purpose to ban TikTok as a means of censoring content. They claim the divestiture exemp- tion cannot be satisfied in the time allowed by the Act, which effectively makes it a ban. Conversely, they argue an exclusion from the definition of “covered company” — for entities that operate an “application whose primary purpose is to allow users to post product reviews, business reviews, or travel information and reviews,” § 2(g)(2)(B) — creates a loophole to the generally applicable track so large that no other company is likely ever to be subjected to the prohibitions of the Act. [10] The upshot, according to TikTok, is that the Congress *46 “purpose-built the Act to ban TikTok because it objects to TikTok’s content.” TikTok Reply Br. 28.
We discern no such motive from the divestiture provisions or the design of the generally applicable framework. Although the Government does not rebut TikTok’s argument that 270 days is not enough time for TikTok to divest given its high degree of integration with ByteDance, 270 days is a substantial amount of time. If TikTok (or any company subject to the Act) is unable to divest within 270 days, it can do so later and thereby lift the prohibitions. § 2(c)(1)(A)–(B). Consequently, we detect no illicit motive on the part of the Congress to ban TikTok and suppress its speech by means of the divestiture provisions.
The same is true of the reviews exclusion, which appears to reflect a good-faith effort by the Congress to narrow the scope of the general track to applications the Congress deter- mined to present the greatest risks to national security. That the Congress created a new mechanism by which the Executive can counter threats similar to TikTok in the future — and excluded a category of applications from that framework — does not suggest the Congress’s national security concerns specific to TikTok were a charade. In fact, the Congress was not required to include a generally applicable framework at all; it could have focused only on TikTok. See Williams-Yulee , 575 U.S. at 452 (“The First Amendment does not put [the Congress] to [an] all-or-nothing choice”). The Congress was entitled to address the threat posed by TikTok directly and create a generally applicable framework, however imperfect, for future use. It would be inappropriate to “punish” the Congress for attempting to address future national security threats by inferring an impermissible motive. Id.
Second, TikTok contends the Government’s content- manipulation rationale is speculative and based upon factual errors. TikTok fails, however, to grapple fully with the Government’s submissions. On the one hand, the Government acknowledges that it lacks specific intelligence that shows the PRC has in the past or is now coercing TikTok into manipulat- ing content in the United States. On the other hand, the Government is aware “that ByteDance and TikTok Global have taken action in response to PRC demands to censor content outside of China.” The Government concludes that ByteDance and its TikTok entities “have a demonstrated history of manipulating the content on their platforms, including at the direction of the PRC.” Notably, TikTok never squarely denies that it has ever manipulated content on the TikTok platform at the direction of the PRC. Its silence on this point is striking given that “the Intelligence Community’s concern is grounded in the actions ByteDance and TikTok have already taken over- seas.” It may be that the PRC has not yet done so in the United States or, as the Government suggests, the Government’s lack of evidence to that effect may simply reflect limitations on its ability to monitor TikTok.
In any event, the Government reasonably predicts that TikTok “would try to comply if the PRC asked for specific actions to be taken to manipulate content for censorship, propa- ganda, or other malign purposes” in the United States. That conclusion rests on more than mere speculation. It is the Government’s “informed judgment” to which we give great weight in this context, even in the absence of “concrete evi- dence” on the likelihood of PRC-directed censorship of TikTok in the United States. Humanitarian Law Project , 561 U.S. at 34–35.
The purported factual errors identified by TikTok do not alter that conclusion. TikTok principally faults the Government *48 for claiming the recommendation engine is “based in China” because it now resides in the Oracle cloud. TikTok Reply Br. 21–22. No doubt, but the Government’s characterization is nonetheless consistent with TikTok’s own declarations. TikTok’s declarants explained that now and under its proposed NSA “ByteDance will remain completely in control of developing the Source Code for all components that comprise ‘TikTok’ . . . including the Recommendation Engine.” They likewise represent that TikTok presently “relies on the support of employees of other ByteDance subsidiaries” for code development. Even when TikTok’s voluntary mitigation measures have been fully implemented, the “source code supporting the TikTok platform, including the recommenda- tion engine, will continue to be developed and maintained by ByteDance subsidiary employees, including in the United States and in China.” TikTok is therefore correct to say the recommendation engine “is stored in the Oracle cloud,” but gains nothing by flyspecking the Government’s characteriza- tion of the recommendation engine still being in China.
b. The Act is narrowly tailored.
The TikTok-specific provisions of the Act are narrowly
tailored to further the Government’s two national security
interests. “It bears emphasis that, under the strict-scrutiny
standard, a restriction must be narrowly tailored, not perfectly
tailored.”
In re Sealed Case
,
(i) TikTok’s proposed NSA TikTok presents its proposed NSA as a less restrictive alternative. TikTok contends that, at minimum, our considera- tion of this alternative implicates factual disputes that require additional proceedings. TikTok, however, misapprehends the thrust of the Government’s objection to the proposed NSA. A senior Executive Branch official involved in the negotiations provided several reasons for which the Executive rejected the proposal. These included lack of U.S. visibility into PRC activity, the Executive’s inability to monitor compliance with the NSA, and therefore its inadequate ability to deter non- compliance; insufficient operational independence for TikTok; and insufficient data protections for Americans. Moreover, and “most fundamentally,” the NSA “still permitted certain data of U.S. users to flow to China, still permitted ByteDance execu- tives to exert leadership control and direction over TikTok’s US operations, and still contemplated extensive contacts between the executives responsible for the TikTok U.S. plat- form and ByteDance leadership overseas.” At bottom, acceptance of “the Final Proposed NSA would ultimately have relied on the Executive Branch trusting ByteDance” to comply with the agreement, which the Government understandably judged it could not do. Based upon this array of problems, the Executive rejected the proposal and pursued a legislative solution.
TikTok adamantly disagrees with the Executive’s judg- ment. It is not, however, the job of the petitioners or of the *50 courts to substitute their judgments for those of the political branches on questions of national security. See Hernández v. Mesa , 589 U.S. 93, 113 (2020). Understandably, TikTok therefore attempts to couch its disagreement in factual terms. But TikTok does not present any truly material dispute of fact.
Consider, for example, TikTok’s claim that data anonymization under TikTok’s proposed NSA would effec- tively mitigate the Government’s concerns. The Government does not dispute that TikTok’s proposal provides for data anonymization; rather, it deems this protection vulnerable to circumvention and therefore insufficient to resolve the Government’s data-related concerns. That is a dispute of judg- ment not of fact. A similar point applies to the parties’ disagree- ment regarding the feasibility of Oracle reviewing TikTok’s source code for the Government. TikTok’s declarant says Oracle could apply methods consistent with industry standards to streamline that review and points out that TikTok’s proposed NSA would require Oracle to conduct its initial review in 180 days. The Government does not disagree; rather, it doubts the adequacy of Oracle’s review of the source code — notwithstanding “Oracle’s considerable resources” — based upon extensive technical conversations with Oracle. Moreover, even after “assuming every line of Source Code could be monitored and verified,” the Government still concluded that “the PRC could exert malign influence” through commercial features of the platform that would not be identified through a review of the code. TikTok’s disagreement with the Government boils down to a dispute about the sufficiency of Oracle’s review to mitigate threats posed by the PRC, which is a matter of judgment, not of fact.
The same is true regarding the role of TTUSDS in limiting the PRC’s ability to control TikTok through ByteDance. The Government concludes that TTUSDS would be insufficiently *51 independent of ByteDance, fears TTUSDS could be pressured to do the latter’s bidding, and doubts TTUSDS could prevent interference by ByteDance. Indeed, the Government predicts that “TTUSDS personnel here would not resist demands to comply” with directives “even if aware of pressure from the PRC government.” Whether TTUSDS sufficiently mitigates the risk of PRC interference through ByteDance is ultimately an issue of judgment, not of fact.
Similarly, the parties’ dispute about the adequacy of the temporary shutdown option — or “kill switch” — under the NSA centers on the Government’s ultimate conclusion regarding the sufficiency of that option. The Government’s declarant on this point explains that the “temporary stop would not . . . give the U.S. Government anything resembling com- plete discretion to shut down the TikTok platform based on its own independent assessment of national security risk and assessments from the U.S. Intelligence Community.” TikTok’s declarant, by contrast, characterizes the so-called “kill switch” as a “unilateral remedy” of unparalleled “magnitude in a CFIUS mitigation agreement,” which could be applied by the Government if TikTok deployed unreviewed source code or if TikTok violates the protocols for handling Protected Data. Rhetoric aside, the substance of TikTok’s objection is the Government’s ultimate conclusion that the shutdown option would not adequately address the Government’s concerns because of the limited scope of the shutdown option as well as the Government’s inability to monitor TikTok.
In sum, even if we resolved every supposed factual dispute
in TikTok’s favor, the result would be the same. For us to
conclude the proposed NSA is an equally or almost equally
effective but less restrictive alternative, we would have to reject
the Government’s risk assessment and override its ultimate
judgment. That would be wholly inappropriate after Executive
*52
Branch officials “conducted dozens of meetings,” considered
“scores of drafts of proposed mitigation terms,” and engaged
with TikTok as well as Oracle for more than two years in an
effort to work out an acceptable agreement. Here “respect for
the Government’s conclusions is appropriate.”
Humanitarian
Law Project
,
The petitioners attempt to draw a distinction between the Executive’s rejection of the proposed NSA and the Congress’s deliberations prior to passing the Act. The petitioners complain the Congress failed even to consider TikTok’s proposed NSA. Because the Act applies narrowly to the TikTok platform, TikTok goes so far as to argue the Congress was required to make legislative findings to explain its rationale for passing the Act. These objections are unavailing. The Congress “is not obligated, when enacting its statutes, to make a record of the type that an administrative agency or court does to accommo- date judicial review.” Time Warner Entm’t Co. v. FCC , 93 F.3d 957, 976 (D.C. Cir. 1996) (cleaned up); Sable Commc’ns of Cal., Inc. v. FCC , 492 U.S. 115, 133 (1989) (Scalia, J., concurring) (“Neither due process nor the First Amendment requires legislation to be supported by committee reports, floor debates, or even consideration, but only by a vote”). Moreover, the petitioners cannot credibly claim the Congress was any less aware than the Executive of the proposed NSA as a potential alternative. Prior to passage of the Act, while the Executive was negotiating the proposed NSA with TikTok, Executive Branch officials briefed congressional committees several times. The record shows that congresspersons were aware of TikTok’s voluntary mitigation efforts; TikTok and its supporters, includ- ing the PRC itself, lobbied the Congress not to pass the Act; and TikTok displayed “a pop-up message urging users to contact their representatives about the Act,” which prompted a deluge of calls to congresspersons. We think it clear the *53 Congress did not reject the proposed NSA for lack of familiar- ity; like the Executive, the Congress found it wanting.
To qualify as a less restrictive alternative, the proposed
NSA must “accomplish the Government’s goals equally or
almost equally effectively.”
In re Sealed Case
,
(ii)
Other options
The petitioners suggest a variety of other options that the
Government also found inadequate. These include disclosure
or reporting requirements, the Government using speech of its
own to counter any alleged foreign propaganda, limiting
TikTok’s collection of location and contact data, and extending
the ban of TikTok on government devices to government
employees’ personal devices. None would “accomplish the
Government’s goals equally or almost equally effectively.”
In
re Sealed Case
,
The first two suggestions obviously fall short. As the Government points out, covert manipulation of content is not a type of harm that can be remedied by disclosure. The idea that the Government can simply use speech of its own to counter the risk of content manipulation by the PRC is likewise naïve. Moreover, the petitioners’ attempt to frame the use of Government speech as a means of countering “alleged foreign propaganda,” Creator Br. 54, is beside the point. It is the “secret manipulation of the content” on TikTok — not foreign propa- ganda — that “poses a grave threat to national security.” Gov’t Br. 36. No amount of Government speech can mitigate that threat nearly as effectively as divestiture.
The petitioners’ other proposals are similarly flawed. Creators’ contention that the Government “could simply ban TikTok from collecting . . . location and contact data” fundamentally misapprehends the Government’s data- collection concerns, which are not limited to two types of data. Creator Reply Br. 29. The data-collection risks identified by the Government include the PRC’s ability to use TikTok for “bulk collection of data” and for “targeted collection on individuals.” Gov’t Br. 48. Indeed, the FBI has specifically assessed that “TikTok could facilitate the PRC’s access to U.S. users’ data, which could enable PRC espionage, technology transfer, data collection and influence activities.” For example, the PRC could use TikTok data to enhance its “artificial intelli- gence capabilities” and obtain “extensive information about users and non-users, including U.S. Government and U.S. intelligence community employees, U.S. political dissidents, and other individuals of interest to the PRC.” Moreover, even if the Government’s concerns were limited to certain categories of data, its inability to monitor TikTok makes a targeted prohibition on the collection of specific types of data less effective than divestiture.
For similar reasons, a limited prohibition addressing government employees would not suffice. The Government’s concern extends beyond federal employees to “family mem- bers or potential future government employees (many of whom may be teenagers today, a particular problem given TikTok’s popularity among young people).” Indeed, as the Government emphasizes, the Congress was legislating “in the interest of all Americans’ data security.” Gov’t Br. 58. A more limited prohibition would not be as effective as divestiture.
The User Petitioners also identify as options various legislative proposals, such as the Adversarial Platform Prevention Act of 2021, S. 47, 117th Cong. (2021); Internet Application I.D. Act, H.R. 4000, 117th Cong. (2021); and the TELL Act, H.R. 742, 118th Cong. (2023), that the Congress did not adopt. In substance, these proposals are similar to the alternatives we just considered and found less effective than divestiture. If anything, those unenacted lesser legislative pro- posals undermine rather than advance the User Petitioners’ preferred alternatives: That the Congress considered a series of other measures before ultimately adopting the Act implies only that the Congress determined nothing short of divestiture would sufficiently avoid the risks posed by TikTok.
In short, the petitioners suggest an array of options none of which comes close to serving either, much less both, the Government’s goals as effectively as does divestiture. Each consequently fails to qualify as a less restrictive alternative for purposes of the First Amendment.
(iii) Overinclusive / underinclusive The petitioners contend the Act is both overinclusive and underinclusive. They argue the Act is overinclusive primarily because the TikTok-specific provisions apply to another ByteDance product, CapCut, that can be used to edit videos on *56 various platforms including TikTok but does not collect user data or present an opportunity for PRC manipulation of con- tent. Given the Government’s well-supported concerns about ByteDance, it was necessary for the Act to apply to all ByteDance entities. Moreover, the petitioners fail to demon- strate that neither of the Government’s two national security concerns implicate CapCut. We therefore conclude the TikTok-specific provisions of the Act are not overinclusive.
We likewise conclude the Act is not fatally underinclusive.
The main purpose of inquiring into underinclusiveness is “to
ensure that the proffered state interest actually underlies the
law.”
Nat’l Ass’n of Mfrs. v. Taylor
,
* * * To summarize our First Amendment analysis: The Government has provided two national security justifications for the Act. We assumed without deciding the Act is subject to strict scrutiny and we now uphold the TikTok-specific portions of the Act under each justification. This conclusion is sup- ported by ample evidence that the Act is the least restrictive means of advancing the Government’s compelling national security interests.
C. Equal Protection
TikTok argues that the Act violates its right to the equal protection of the laws because it singles out TikTok for disfa- vored treatment relative to other similarly situated platforms. The Government contends its justifications for the Act satisfy the requirement of equal protection and add that TikTok received more process than a company would receive under the generally applicable provisions. We conclude the Act is con- sistent with the requirement of equal protection .
“In equal protection challenges the critical question is always whether there is an appropriate governmental interest suitably furthered by the differential treatment at issue.” Cmty- Serv. Broad. of Mid-Am., Inc. v. FCC , 593 F.2d 1102, 1122 (D.C. Cir. 1978) (cleaned up). This question “lies at the intersection” of equal protection and the First Amendment. News Am. Pub., Inc. v. FCC , 844 F.2d 800, 804 (D.C. Cir. 1988) (cleaned up).
Although we review “conventional economic legislation”
under a “minimum rationality” standard,
id.
at 802, we have
held something “more is required than ‘minimum rationality’”
when a regulation burdens “a single publisher/broadcaster,”
id.
at 814.
See also BellSouth I
, 144 F.3d at 68 (explaining that
*58
News America
does not require strict scrutiny for “statutes
singling out particular persons for speech restrictions”);
Cmty-
Serv. Broad. of Mid-Am., Inc.
,
TikTok’s equal protection argument boils down to point- ing out that TikTok alone is singled out by name in the Act, unlike companies that in the future may be subject to the gener- ally applicable provisions of the Act. Merely singling a com- pany out, however, does not amount to an equal protection violation if doing so furthers an appropriate governmental interest. The controlling question is “whether there is an appropriate governmental interest suitably furthered by the differential treatment at issue.” Cmty-Serv. Broad. of Mid-Am., Inc. , 593 F.2d at 1122–23 (holding statute violated First and Fifth Amendments by unjustifiably burdening only non- commercial broadcasters). Here the Government justified the Act by presenting two national security risks specific to the TikTok platform. By naming TikTok in the Act, the Congress ensured TikTok-related risks were addressed promptly. Simultaneously creating a generally applicable framework gave the Executive a tool to address similar risks that may come to light in the future. This differential treatment furthers the Government’s national security interest in countering the immediate threat posed by the PRC’s control of TikTok.
The governmental interests here also stand in stark con-
trast to the case upon which TikTok primarily relies, in which
the “sole apparent difference” in treatment between similarly
situated broadcasters was due to “an accident of timing.”
News
*59
Am. Pub., Inc.
,
Moreover, as the Government notes, in certain respects TikTok received more process than would a company coming under the generally applicable provisions. TikTok participated in a prolonged negotiation with the Executive that featured numerous meetings and several proposals. It also received individualized consideration by the Congress prior to being required to divest. In contrast, under the generally applicable provisions the Executive need only provide “public notice” and issue a “public report” to the Congress prior to requiring a com- pany to sever its ties to an adversary nation. § 2(g)(3)(B). In short, the Act singled out TikTok because of its known characteristics and history. It therefore did not violate TikTok’s constitutional right to equal protection of the laws.
D. The Bill of Attainder Clause
TikTok next claims the Act is a bill of attainder, and there- fore prohibited by Article I, § 9, clause 3 of the Constitution. The Government responds that the Bill of Attainder Clause does not apply to corporations and that, in any event, the Act does not constitute a legislative punishment. We agree that the Act is not a bill of attainder.
A law is a bill of attainder if it “(1) applies with specificity,
and (2) imposes punishment.”
BellSouth Corp. v. FCC
(
BellSouth II
),
Before turning to those tests, however, we briefly address the Government’s threshold argument that the Bill of Attainder Clause does not apply to corporations. In other cases, we have assumed without deciding that the clause applies to corpora- tions but emphasized that differences between commercial entities and persons need to be considered. See, e.g. , Kaspersky Lab, Inc. v. DHS , 909 F.3d 446, 453–54, 461–63 (D.C. Cir. 2018) (assuming the Bill of Attainder Clause protects corpora- tions but emphasizing the differences between corporations and “living, breathing human beings”); BellSouth I , 144 F.3d at 63 & n.5 (assuming the clause protects corporations but recognizing the importance of understanding “its effect on flesh-and-blood people”). We take the same approach here.
To determine whether a law constitutes a punishment, we analyze:
(1) whether the challenged statute falls within the historical meaning of legislative punishment [the historical test];
(2) whether the statute, viewed in terms of the type and severity of burdens imposed, reasonably can be said to further nonpunitive legislative purposes [the functional test]; and (3) whether the legislative record evinces a congres- sional intent to punish [the motivational test]. Kaspersky Lab, Inc. , 909 F.3d at 455 (cleaned up). The Act clearly is not a bill of attainder judged by any of these tests.
TikTok contends the Act satisfies the historical test because it bars TikTok from its chosen business. TikTok *61 reasons the prohibitions of the Act are close analogs to two categories of legislative action historically regarded as bills of attainder: confiscation of property and legislative bars to participation in a specific employment or profession. See BellSouth II , 162 F.3d at 685 (explaining the historical understanding of punishment). According to TikTok, the Act effectively requires TikTok to relinquish its property or see it rendered useless, and it precludes TikTok from continuing to participate in a legitimate business enterprise. As already explained, however, the Act requires a divestiture — that is, a sale, not a confiscation — as a condition of continuing to operate in the United States. See BellSouth I , 144 F.3d at 65 (explaining that although “structural separation is hardly costless, neither does it remotely approach the disabilities that have traditionally marked forbidden attainders”); see also Kaspersky Lab, Inc. , 909 F.3d at 462–63 (comparing a law requiring the Government to remove from its systems a Russia- based company’s software to the business regulations in the BellSouth cases). Nor is the divestiture requirement analogous to a legislative bar on someone’s participation in a specific employment or profession. See Kaspersky Lab, Inc. , 909 F.3d at 462 (rejecting a similar analogy in part “because human beings and corporate entities are so dissimilar” (cleaned up)).
The closer historical analog to the Act is a line-of-business
restriction, which does not come within the historical meaning
of a legislative punishment.
See BellSouth II
,
The Act also passes muster under the functional test. For purposes of this analysis, the “question is not whether a burden is proportionate to the objective, but rather whether the burden is so disproportionate that it belies any purported nonpunitive goals.” Kaspersky Lab, Inc. , 909 F.3d at 455 (cleaned up). Considering our conclusion that the Act passes heightened scrutiny for purposes of the First Amendment, it obviously satisfies the functional inquiry here: The Act furthers the Government’s nonpunitive objective of limiting the PRC’s ability to threaten U.S. national security through data collection and covert manipulation of information. The Government’s solution to those threats “has the earmarks of a rather conven- tional response to a security risk: remove the risk.” Id. at 457 (cleaned up). In other words, the Government’s attempt to address the risks posed by TikTok reflects a forward-looking prophylactic, not a backward-looking punitive, purpose. That is sufficient to satisfy the functional analysis. See id. at 460 (stating the functional test “does not require that the Congress precisely calibrate the burdens it imposes to the goals it seeks to further or to the threats it seeks to mitigate” (cleaned up)).
The so-called motivational test, for its part, hardly merits
discussion. “Given the obvious restraints on the usefulness of
*63
legislative history,” congressional intent to punish is difficult
to establish.
Id.
at 463 (cleaned up);
see also BellSouth II
,
E. The Takings Clause
TikTok claims the Act constitutes a per se regulatory taking in violation of the Fifth Amendment because it will render TikTok defunct in the United States. The Government counters that TikTok has assets that can be sold, and that the Act requires only divestiture, which need not be uncompen- sated. Although the Act will certainly have a substantial effect on the TikTok platform in the United States, regardless whether TikTok divests, the Act does not qualify as a per se regulatory taking.
The Supreme Court recognizes two situations in which regulatory action constitutes a per se taking: (1) where the government requires that an owner suffer a “physical invasion of [its] property,” and (2) where a regulation “completely deprives an owner of all economically beneficial use of [its] property.” Lingle v. Chevron U.S.A. Inc. , 544 U.S. 528, 538 (2005) (cleaned up); see Cedar Point Nursery v. Hassid , 594 U.S. 139, 153 (2021) (explaining the first category includes temporary invasions of property). TikTok’s argument is of the second variety, but it does not demonstrate the com- plete deprivation such a claim requires.
Here the causal connection between the Act and the alleged diminution of value is attenuated because the Act *64 authorizes a qualified divestiture before (or after) any prohibi- tions take effect. That presents TikTok with a number of possibilities short of total economic deprivation. ByteDance might spin off its global TikTok business, for instance, or it might sell a U.S. subset of the business to a qualified buyer.
TikTok dismisses divestiture as impractical. One of the main impediments, however, appears to be export prohibitions that the PRC erected to make a forced divestiture more difficult if not impossible. But the PRC, not the divestiture offramp in the Act, is the source of TikTok’s difficulty. TikTok would have us turn the Takings Clause into a means by which a for- eign adversary nation may render unconstitutional legislation designed to counter the national security threats presented by that very nation.
In any event, TikTok has not been subjected to a complete deprivation of economic value. Beyond characterizing divesti- ture as impossible, TikTok does not dispute that it has assets that can be sold apart from the recommendation engine, includ- ing its codebase; large user base, brand value, and goodwill; and property owned by TikTok. In other words, TikTok has several economically beneficial options notwithstanding the PRC’s export restriction.
F. Alternative Relief
As an alternative to permanently enjoining the Act, the petitioners suggest we issue a temporary injunction and appoint a special master to make procedural recommendations or recommend factual findings. Because we have now resolved the case on the merits, we deny these requests as moot. The petitioners further object to the Government having filed classified material and releasing to them only a redacted ver- sion. Our decision, however, rests solely on the unredacted, public filings in this case. See China Telecom (Ams.) Corp. , *65 57 F.4th at 264 (similarly relying on an unclassified record). Notwithstanding the significant effect the Act may have on the viability of the TikTok platform, we conclude the Act is valid based upon the public record. [11]
III. Conclusion
We recognize that this decision has significant implica- tions for TikTok and its users. Unless TikTok executes a qualified divestiture by January 19, 2025 — or the President grants a 90-day extension based upon progress towards a qualified divestiture, § 2(a)(3) — its platform will effectively be unavailable in the United States, at least for a time. Consequently, TikTok’s millions of users will need to find alternative media of communication. That burden is attributa- ble to the PRC’s hybrid commercial threat to U.S. national security, not to the U.S. Government, which engaged with TikTok through a multi-year process in an effort to find an alternative solution.
The First Amendment exists to protect free speech in the United States. Here the Government acted solely to protect that freedom from a foreign adversary nation and to limit that adversary’s ability to gather data on people in the United States.
For these reasons the petitions are, Denied.
S RINIVASAN , Chief Judge , concurring in part and concurring in the judgment:
I fully join all aspects of the court’s opinion today other than Part II.B, which rejects TikTok’s First Amendment challenge. As to that challenge, I agree with my colleagues that the Act does not violate the First Amendment. But I reach that conclusion via an alternate path. My colleagues do not decide whether the Act should be subjected to the strictest First Amendment scrutiny or instead the lesser standard of intermediate scrutiny because, in their view, the Act satisfies strict scrutiny regardless. I see no need to decide whether the Act can survive strict scrutiny, because, in my view, the Act need only satisfy intermediate scrutiny, which it does. I would thus answer the question my colleagues leave open while leaving open the question they answer.
Two features of the Act support applying intermediate rather than strict scrutiny to resolve TikTok’s First Amendment challenge. First, in step with longstanding restrictions on foreign control of mass communications channels, the activity centrally addressed by the Act’s divestment mandate is that of a foreign nation rather than a domestic speaker—indeed, not just a foreign nation but a designated foreign adversary. Second, the Act mandates divestment of that foreign adversary’s control over TikTok for reasons lying outside the First Amendment’s heartland: one reason that is wholly unrelated to speech, and another that, while connected to speech, does not target communication of any specific message, viewpoint, or content.
In those circumstances, the Act’s divestment mandate need not be the least restrictive means of achieving its national- security objectives, as strict scrutiny would require. Rather, it is enough if, per intermediate scrutiny, the divestment mandate is not substantially broader than necessary to meet those goals. The Act meets that standard.
A.
TikTok’s First Amendment challenge “implicates the gravest and most delicate duty that courts are called on to perform: invalidation of an Act of Congress.” Hodge v. Talkin , 799 F.3d 1145, 1157 (D.C. Cir. 2015) (formatting modified) (quoting Blodgett v. Holden , 275 U.S. 142, 147–48 (1927) (Holmes, J., concurring)). And that “most delicate duty” presents itself here in a setting in which courts already proceed with suitable caution—when called upon to review the political branches’ judgments about national security. A strong bipartisan majority of both Houses of Congress, together with two successive Presidents (one of whom is also the President- elect), have determined that divesting TikTok from PRC control is a national-security imperative. See Op., ante , at pp. 11–16.
While that is the political branches’ across-the-board
assessment of a pressing national-security issue today, we also
take stock of history when considering whether their response
stays within the bounds of the First Amendment. An
established “history and tradition of regulation [is] relevant
when considering the scope of the First Amendment.”
City of
Austin v. Reagan Nat’l Advert. of Austin, LLC
,
The first communications medium capable of reaching mass audiences in real time—radio—was subject to restrictions on foreign ownership and control from the very outset. The Radio Act of 1912 required radio operators engaged in interstate (or international) communications to obtain a license from the Secretary of Commerce and Labor, but Congress made licenses available only to U.S. citizens or companies. Pub. L. No. 62-264, §§ 1–2, 37 Stat. 302, 302–03 (repealed 1927). Congress then extended the restrictions to encompass foreign control (not just foreign ownership) in the Radio Act of 1927, prohibiting licensing of any company if it had a foreign officer or director or if one-fifth of its capital stock was in foreign hands. Pub. L. No. 69-632, § 12, 44 Stat. 1162, 1167 (repealed 1934).
Within a few years, the Communications Act of 1934, Pub. L. No. 73-416, 48 Stat. 1064, shored up the restrictions on foreign control. Section 310 of the law incorporated with little change the 1927 Act’s foreign-control requirements, and also gave the newly created Federal Communications Commission (FCC) authority to withhold a license if a company is “directly or indirectly controlled” by a foreign-dominated parent company. Id . § 310(a), 48 Stat. at 1086 (emphasis added) (today codified at 47 U.S.C. § 310(b)(4) (2024)). In urging Congress to adopt the additional restrictions on foreign control, the Navy conveyed its concerns that foreign-controlled stations could “be employed in espionage work and in the dissemination of subversive propaganda.” Hearings on H.R. 8301 Before the H. Comm. on Interstate & Foreign Com., 73d Cong. 26 (1934). The FCC has described Section 310’s original purpose as “protect[ing] the integrity of ship-to-shore and governmental communications” from foreign interference and “thwart[ing] the airing of foreign propaganda on broadcast stations.” Foreign Investment in Broadcast Licenses, 78 Fed. Reg. 75563, 75564 (Nov. 13, 2013).
Section 310 continues to restrict foreign control of radio licenses, including ones used for broadcast communication and wireless cellular services. See 47 U.S.C. § 310(a)–(b). And while that provision regulates wireless licenses, limitations on foreign control also exist for wired transmission lines under Section 214 of the same law. 47 U.S.C. § 214(a); see also id. § 153(11), (50)–(53).
When deciding whether to issue or revoke a Section 214
authorization, the FCC considers “the public convenience and
necessity,”
id.
§ 214(c), including the implications for
“national defense,”
id.
§ 151. In conducting that inquiry, the
FCC assesses whether direct or indirect foreign ownership or
control of a transmission line raises national-security or
foreign-policy concerns.
See Rules & Policies on Foreign
Participation in the U.S. Telecomm. Mkt.
, 12 FCC Rcd. 23891,
23918–21 (1997). The FCC consults with Executive Branch
agencies “to help assess national security and other concerns
that might arise from a carrier’s foreign ownership.”
China
Telecom (Americas) Corp. v. FCC
,
Notably, the FCC in recent years has exercised its Section 214 authority to deny or revoke transmission authorizations in the case of U.S. entities subject to ultimate Chinese control. The Commission’s rationale has mirrored Congress’s motivation for the Act we consider in this case—i.e., national- security concerns that the PRC could leverage its control over foreign parent companies to require U.S. subsidiaries to provide China with access to U.S. communications lines, thereby enabling espionage and other harmful undertakings. See Pac. Networks Corp. & ComNet (USA) LLC , 37 FCC Rcd. 4220 (2022); China Telecom (Americas) Corp. , 36 FCC Rcd.
5
15966 (2021);
China Mobile Int’l (USA)
, 34 FCC Rcd. 3361
(2019). This court has affirmed those FCC decisions.
See Pac.
Networks Corp. v. FCC
,
China Telecom , for example, involved a U.S. company with a Section 214 authorization whose parent corporation was majority-owned by a Chinese governmental entity. See F.4th at 260, 265. The FCC’s revocation of China Telecom’s authorization was “grounded [in] its conclusion that China Telecom poses an unacceptable security risk” because “the Chinese government is able to exert significant influence over [it].” Id. at 265. In rejecting China Telecom’s claim that the asserted national-security risk was unduly speculative, we noted that Chinese law obligates Chinese companies “to cooperate with state-directed cybersecurity supervision and inspection,” and we cited “compelling evidence that the Chinese government may use Chinese information technology firms as vectors of espionage and sabotage.” Id. at 265–66. We additionally explained that “[i]n the national security context,” the FCC “need not wait for a risk to materialize before revoking a section 214 authorization.” Id. at 266.
China Telecom
is a present-day application of the kinds of
restrictions on foreign control that have existed in the
communications arena since the dawn of radio. That
longstanding regulatory history bears on the First Amendment
analysis here.
See City of Austin
,
To be sure, because communications media reaching mass
audiences in real time “were not present in the founding era,”
the regulatory history naturally does not date back that far.
See
City of Austin
,
So, in City of Austin , the Supreme Court recently assessed which of those standards should govern a challenge to a law attaching different restrictions to off-premises and on-premises signage. See id. at 67–69. The Court explained that comparable regulations emerged relatively soon after outdoor billboards first appeared in the 1800s. See id. at 65–66, 75. To the Court, that “unbroken tradition of on-/off-premises distinctions counsel[ed] against” subjecting the challenged law to strict scrutiny. Id. at 75. If so there, so too here.
B.
In
City of Austin
, the Supreme Court considered the
longstanding regulatory history as part of its inquiry into
whether the law in question should be deemed content based or
content neutral.
See
There can also be, though, an antecedent question:
whether the First Amendment applies at all. The question
arises here because the effect of the Act’s divestment mandate
falls most directly on foreign entities: the Act targets the PRC,
a foreign sovereign, and the divestment mechanism established
by Congress necessarily encompasses ByteDance, a foreign
company subject to the PRC’s control. That recognition brings
into play the settled understanding that “foreign organizations
operating abroad have no First Amendment rights.”
Agency for
Int’l Dev. v. All. for Open Soc’y Int’l Inc.
,
The Act requires TikTok to divest the corporate parent,
ByteDance, because ByteDance is subject to the PRC’s
control. ByteDance developed and maintains the source code
underlying TikTok’s recommendation engine,
see
Simkins
Decl. ¶¶ 52, 57, 90 (TikTok App. 738, 740, 751); Presser Decl.
¶¶ 63–64 (TikTok App. 832), so the company has the ability to
curate the content sent to TikTok users. That kind of curation
function, when the First Amendment applies, is protected
expressive activity. As the Supreme Court recently explained,
“presenting a curated compilation of speech originally created
by others” via a social media app is a form of
expression.
Moody v. NetChoice, LLC
,
To the extent the PRC or ByteDance might wish to adjust
the content viewed by U.S. users of TikTok, those curation
decisions would be made abroad.
See
Milch Decl. ¶ 29
(TikTok App. 661) (explaining that TikTok’s proposed
security measures contemplate “continued reliance on
ByteDance engineers for . . . its recommendation engine”).
The PRC and ByteDance thus would lack any First
Amendment rights in connection with any such curation
actions.
Agency for Int’l Dev.
,
That case involved foreign organizations’ speech that was
targeted in part at the United States, yet the Court still applied
the rule that the foreign speakers lack any First Amendment
rights when engaged in expressive activity abroad. The federal
statute challenged in
Agency for International Development
required organizations receiving certain U.S. aid dollars to
espouse a policy opposing prostitution.
Id.
at 432. The Court
first held that the compelled adoption of an anti-prostitution
viewpoint violated the First Amendment as applied to U.S.
funding recipients.
See Agency for Int’l Dev. v. All. for Open
Soc’y Int’l Inc.
,
In short, while the Act’s divestment mandate directly affects—and aims to eliminate—the ability of the PRC and ByteDance to engage with U.S. users of a PRC-controlled TikTok, it raises no First Amendment concerns vis-à-vis those foreign actors.
C.
Even if ByteDance and the PRC lack First Amendment rights to assert against the Act’s divestment mandate, what about the U.S.-based petitioners’ free-speech claims? The principal U.S. petitioners are: (i) TikTok Inc., the U.S. subsidiary of ByteDance that provides the TikTok platform in the United States; and (ii) U.S. TikTok users, who are both creators and viewers of TikTok content.
1.
For TikTok Inc., the Act is designed to sever ByteDance from the platform but leave untouched TikTok Inc.’s expression on a post-divestment version of the app. TikTok Inc. both creates and curates content on the platform, and the Act does not restrict those speech and curation choices. TikTok Inc. posts videos to its own TikTok account and would remain fully free to continue doing so post-divestment. The company can also engage in content moderation, including through enforcement of community guidelines that excise videos containing nudity, for instance. See Op., ante , at p. 27. To the extent those choices are TikTok Inc.’s own, the company could maintain the same editorial policies on a post- divestment version of the app.
TikTok also claims that TikTok Inc.’s deployment of the platform’s recommendation engine in the U.S. is itself an expressive decision. Even assuming so, after divestment, a non-Chinese-controlled TikTok could still use the same algorithm to promote the same exact mix of content presently appearing on the app. According to TikTok, however, Chinese law would prevent the export of the algorithm fueling the recommendation engine without the PRC’s approval, which it would not grant. TikTok Br. 24. The Act , though, would not dictate that outcome. Rather, the PRC, backed by Chinese law, would. And Congress of course need not legislate around another country’s preferences to exercise its own powers constitutionally—much less the preferences of a designated foreign adversary, the very adversary whom Congress determined poses the fundamental threat to national security prompting the Act in the first place.
2.
The last group of petitioners bringing a First Amendment claim are users who create and consume content on the TikTok platform. They face the prospect of the app becoming unavailable to them if a divestment does not occur within the window allowed by Congress, or of an app potentially altered in certain ways if a divestment were to take place.
A threshold question bearing significantly on the assessment of their First Amendment challenge is which standard of scrutiny should apply: strict or intermediate scrutiny. The choice can be an important, potentially outcome- determinative one, which is why the Supreme Court can devote entire decisions to the issue. See, e.g. , City of Austin , 596 U.S. 61. That choice here, as is often the case, turns in significant measure on the rationale for the challenged law, which informs whether the law is considered content based or content neutral.
As my colleagues explain, the Act’s divestment mandate rests on two justifications, both of which concern the PRC’s ability (through its control over ByteDance) to exploit the TikTok platform in ways inimical to U.S. national security. See Op., ante, at p. 33. First, the PRC could harvest abundant amounts of information about the 170 million U.S. app users and potentially even their contacts. Second, the PRC could direct the TikTok platform to covertly manipulate the content flowing to U.S. users. To the government, a foreign adversary’s ability to acquire sensitive information on Americans and secretly shape the content fed to Americans would pose a substantial threat to U.S. national security.
Those dual interests are manifested in the terms of the Act, in its central provisions establishing the divestment requirement. The Act defines a “qualified divestiture” as one that removes any ongoing relationship with the foreign adversary-controlled entities with which the app was previously affiliated, including in particular “any cooperation with respect to the operation of a content recommendation algorithm or an agreement with respect to data sharing .” § 2(g)(6)(B) (emphasis added). In the central operative provision of the Act, then, Congress established that a divestiture must satisfy the two national-security concerns invoked by the government in this case: data protection and content manipulation.
An examination of those interests, separately and in combination, shows that the Act does not raise the kinds of core free-speech concerns warranting the application of strict scrutiny. Instead, intermediate scrutiny should apply.
a.
The data-protection rationale is plainly content neutral, supporting the application of intermediate rather than strict scrutiny. There is no sense in which the data-protection interest relates to the content of speech appearing on TikTok. In fact, the interest does not relate to speech at all, raising the question whether it would even trigger intermediate scrutiny if it stood alone.
In
Arcara v. Cloud Books, Inc.
,
Here, similarly, the data-protection rationale has nothing to do with the expressive activity taking place on the TikTok platform. Any enterprise collecting vast amounts of data from users, whatever its line of business, could pose that sort of risk. That is not to diminish the burdens on millions of U.S. users if the TikTok platform were to become unavailable to them as a forum for expressive activity. All of them could be faced with needing to find an alternate venue. The same was true, though, of the bookstore patrons in Arcara , yet the Court still denied *78 the First Amendment challenge to the bookstore’s closure without even applying intermediate scrutiny.
To be sure, the
Arcara
Court observed that First
Amendment scrutiny would apply to a law that “inevitably
single[s] out bookstores or others engaged in First Amendment
protected activities for the imposition of its burden.”
Id.
Even
if that description has salience here—which is not at all clear—
the Court has explained that such a law may be “justified by
some special characteristic” of the regulated entities.
Minneapolis Star v. Minn. Comm’r of Rev.
,
At any rate, there is no need to reach a firm conclusion on whether the data-protection interest, if considered in isolation, would trigger the application of intermediate scrutiny or instead an even more relaxed form of review. That is because the government makes no argument that the Act’s application to TikTok should be sustained based on the data-protection interest alone. It is necessary, then, to engage with the other interest underpinning the Act, to which I turn next.
b.
Congress’s interest in preventing the PRC’s use of TikTok to engage in covert content manipulation is self-evidently connected to speech: it centers on the potential reactions of American viewers to covert content-curation decisions made by the PRC. Still, that interest does not raise heartland First Amendment concerns about content-based restrictions for reasons I will explain—so much so that, even if that interest were the sole rationale for the Act, there would still be a strong argument for applying intermediate rather than strict scrutiny.
It is important to keep in mind, though, that Congress’s covert-content-manipulation concern does not stand alone. There is also its distinct data-protection interest that supports applying (at most) intermediate scrutiny, along with the consistent regulatory history of restricting foreign control of mass communications channels that likewise weighs in favor of intermediate scrutiny. So, the question ultimately is not whether the covert-content-manipulation concern itself would occasion applying strict scrutiny, but rather whether it so strongly and clearly does that it overcomes the other important considerations counseling against strict scrutiny. I believe it does not.
First, even assuming the covert-content-manipulation
concern may bear the indicia of a content-based rationale, it
would do so only marginally. The Supreme Court has used
slightly varying formulations when describing what makes a
law content based, but this recent articulation captures the gist:
not just “
any
examination of speech or expression inherently”
makes a regulation content based; rather, “it is regulations that
discriminate based on ‘the topic discussed or the idea or
message expressed’ that are content based.”
City of Austin
, 596
U.S. at 73–74 (quoting
Reed
,
Congress’s concern about the PRC’s capacity to conduct covert content manipulation on the TikTok platform does not “discriminate based on the topic discussed or the idea or message expressed.” City of Austin , 596 U.S. at 73–74 (internal quotation marks omitted) . Congress desires to prevent the PRC’s secret curation of content flowing to U.S. users regardless of the topic, idea, or message conveyed. See Gov’t Br. 66–68. To be sure, Congress would have concerns about the PRC covertly compelling ByteDance to flood the feeds of American users with pro-China propaganda. But *80 Congress would also have concerns about the PRC sowing discord in the United States by promoting videos—perhaps even primarily truthful ones—about a hot-button issue having nothing to do with China. Indeed, because the concern is with the PRC’s manipulation of the app to advance China’s interests —not China’s views—one can imagine situations in which it would even serve the PRC’s interests to augment anti - China, pro- U.S. content. Suppose, for instance, the PRC determines that it is in its interest to stir an impression of elevated anti-China sentiment coming from the United States—say, to conjure a justification for actions China would like to take against the United States. That would qualify as covert content manipulation of the kind that concerned Congress and supports the Act’s divestment mandate.
Congress’s concern with covert content manipulation by a
foreign adversary in any direction and on any topic—rather
than on particular messages, subjects, or views—is evident in
the Act’s terms and design.
See City of Renton v. Playtime
Theaters, Inc.
,
As is reflected in the title of the Act—“Protecting
Americans From Foreign Adversary Controlled Applications
Act”—Congress aimed not to address specific content but to
address specific actors: in particular, to prevent a “foreign
adversary” from exercising control over covered applications.
In that sense, the law operates in the nature of a speaker-based
*81
restriction. As applied here, what matters is whether a
particular potential curator, the PRC, has the ability to control
(covertly) the content fed to TikTok’s U.S. users, regardless of
what the content may be. True, “laws favoring some speakers
over others demand strict scrutiny” when the “speaker
preference reflects a content preference.”
Reed
, 576 U.S. at
171 (quoting
Turner I
,
In certain respects, in fact, the Act resembles a time, place,
or manner regulation—a type of regulation generally subject to
intermediate scrutiny.
Clark v. Cmty. for Creative Non-
Violence
, 468 U.S. 288, 293 (1984);
Ward v. Rock Against
Racism
,
Those circumstances are far removed from
Lamont v.
Postmaster General
,
This case does not involve the “narrow ground” on which the Court rooted its decision in Lamont : an affirmative obligation to out oneself to the government in order to receive communications from a foreign country that are otherwise permitted to be here. Moreover, whereas this case, as explained, addresses what amounts to a speaker-based regulation without a content preference underpinning it, the law in Lamont drew a viewpoint-based distinction based on whether the government deemed mailed material “communist political propaganda.” Finally, Lamont was not a case about covert content manipulation, the concern driving the Act’s divestment mandate. In that regard, while counterspeech is an available response in the case of a publication designated as “communist political propaganda,” counterspeech is elusive in response to covert (and thus presumably undetected) manipulation of a social media platform.
* * *
For all those reasons, Congress’s concern with the PRC’s potential exercise of covert content manipulation should not give rise to strict scrutiny. That concern does not bear the hallmarks of a content-based rationale; the Act’s other justification concerning data protection is plainly a content- neutral one; and there has been a long regulatory history of restrictions on foreign control of mass communications channels.
D.
To satisfy intermediate scrutiny, a law needs to meet two
requirements: (i) the law must further “important” (or
“substantial” or “legitimate”) governmental interests; and (ii)
the means must be narrowly tailored to serve those interests.
See Turner I
,
Here, the Act satisfies both prongs of the intermediate scrutiny test.
1.
Recall that, as manifested in the Act’s terms and design, see § 2(g)(6)(B), Congress mandated TikTok’s divestment in order to prevent the PRC from capturing the personal data of millions of Americans and surreptitiously manipulating the content the app serves them. Each of those objectives qualifies as an important governmental interest.
a.
The data-protection interest aims to protect U.S. national security by depriving the PRC of access to a vast dataset of granular information on 170 million Americans. Congress’s interest is important and well grounded.
As TikTok does not dispute, the platform collects vast amounts of information from and about its American users. See TikTok App. 820; Privacy Policy, TikTok (Aug. 28, 2024), https://perma.cc/XE6G-F86Q. The government’s national- security concerns about the PRC’s access to that data take two forms. First, the PRC could exploit sensitive data on individual Americans to undermine U.S. interests, including by recruiting assets, identifying Americans involved in intelligence, and pressuring and blackmailing our citizens to assist China. Second, the vast information about Americans collected by TikTok amounts to the type of “bulk” dataset that could “greatly enhance” China’s development and use of “artificial intelligence capabilities.” Vorndran Decl. ¶ 32 (Gov’t App. 37).
Those national-security concerns self-evidently qualify as
important. To be sure, the fears must be “real, not merely
conjectural.”
Turner I
,
When applying intermediate scrutiny, a court “must accord substantial deference to the predictive judgments of Congress,” and “[o]ur sole obligation is to assure that, in formulating its judgments, Congress has drawn reasonable inferences based on substantial evidence.” Turner Broad. Sys., Inc. v. FCC , 520 U.S. 180, 195 (1997) ( Turner II ) (internal quotation marks omitted). That bar is cleared here.
In evaluating whether Congress’s national-security concerns are adequately grounded, we can take stock of the Executive Branch’s elaborations as submitted in declarations. See Humanitarian Law Project , 561 U.S. at 33. As my colleagues set out, Op., ante , at pp. 34–36, and as the *85 government explains, Congress’s data-security concern arises against a backdrop of broadscale “overt and covert actions” by the PRC “to undermine U.S. interests,” Blackburn Decl. ¶ 23 (Gov’t App. 8). Collecting data on Americans is a key part of that multi-faceted strategy. See id. ¶¶ 31–33 (Gov’t App. 10– 11). The PRC has engaged in extensive efforts to amass data on Americans for potential use against U.S. interests. Id. ¶ 31 (Gov’t App. 10–11). And the PRC “is rapidly expanding and improving its artificial intelligence and data analytics capabilities for intelligence purposes,” enabling it to exploit access to large datasets in increasingly concerning ways. Id. ¶ 30 (Gov’t App. 10).
“ByteDance and TikTok present powerful platforms” for those purposes. Id. ¶ 36 (Gov’t App. 13). It is a modus operandi of the PRC to surreptitiously access data through its control over companies like ByteDance. While the PRC has sometimes obtained data through aggressive hacking operations, it also attempts to do so by “leverag[ing] access through its relationships with Chinese companies.” Id. ¶ 33 (Gov’t App. 11). Even if the PRC has yet to discernibly act on its potential control over ByteDance’s access to data on American users in particular, Congress did not need to wait for the risk to become realized and the damage to be done before taking action to avert it. See Humanitarian Law Project , 561 U.S. at 34–35; China Telecom , 57 F.4th at 266–67. That is particularly so in light of the PRC’s broader, long-term geopolitical strategy of pre-positioning assets for potential use against U.S. interests at pivotal moments. See Vorndran Decl. ¶ 12 (Gov’t App. 34); Blackburn Decl. ¶ 26 (Gov’t App. 9).
In these circumstances, in short, Congress’s data- protection concern is hardly speculative or inadequately grounded in this murky corner of national security.
b.
The same is true of Congress’s concern about the PRC’s
covert content manipulation. Our duty to accord deference to
Congress’s determinations when applying
intermediate
scrutiny,
Turner II
,
As the government details and petitioners do not dispute, the PRC engages in an aggressive, global campaign of influence operations against U.S. interests, relying heavily on the internet and social-media platforms. Blackburn Decl. ¶¶ 28–29 (Gov’t App. 9–10). Across the globe, the PRC seeks to “promote PRC narratives . . . and counter other countries’ policies that threaten the PRC’s interests.” Id. ¶ 29 (Gov’t App. 10). That includes increasingly pronounced efforts to “mold” America’s “public discourse” and “magnify” our “societal divisions.” Id.
It was reasonable for Congress to infer from the information available to it that, if directed by the PRC to assist in those efforts, ByteDance and TikTok “would try to comply.” Id. ¶ 69 (Gov’t App. 23). The government points to examples of when “the PRC has exerted control over the content shown on other ByteDance-managed apps.” Vorndran Decl. ¶ 33 *87 (Gov’t App. 38). And were the PRC to exert that kind of covert control over the content on TikTok, it would be “difficult—if not impossible—to detect, both by TikTok users and by law enforcement personnel.” Id. ¶ 34 (Gov’t App. 38). In that context, Congress’s concern with preventing the PRC’s covert content manipulation of the platform readily qualifies as an important, well-founded governmental interest.
In resisting that conclusion, petitioners contend that the
covert-content-manipulation rationale cannot be an important
governmental interest because it is “related to the suppression
of free expression.”
NetChoice
,
As an initial matter, insofar as petitioners believe that a law
can
never
satisfy First Amendment scrutiny if it is “related to
the suppression of free expression,” that is incorrect. The
consequence of a law’s being deemed “related to the
suppression of expression” is not that the law is then per se
invalid, but instead that it is then subject to strict rather than
intermediate scrutiny.
See Humanitarian Law Project
, 561
U.S. at 28 (citing
Texas v. Johnson
,
That conclusion is fully consistent with
NetChoice
, as the
laws at issue there were “related to the suppression of
expression” in a way untrue of the Act. In
NetChoice
, two
states enacted laws addressing perceived bias against
conservative viewpoints on large social-media platforms like
YouTube and Facebook. 144 S. Ct. at 2394. The laws
restricted the platforms’ ability to remove, label, or deprioritize
posts or users based on content or viewpoint.
Id.
at 2395–96.
The laws did so, the Supreme Court explained, in pursuit of an
*88
objective “to correct the mix of speech that the major social-
media platforms present,” so as “to advance [the states’] own
vision of ideological balance.”
Id.
at 2407. The Court
explained that such an interest “is very much related to the
suppression of free expression, and it is not valid, let alone
substantial.”
Id.
(citing
Buckley v. Valeo
,
Here, by contrast, the Act is not grounded in any congressional aim to correct a perceived viewpoint imbalance on the TikTok platform by achieving a different ideological mix. Congress, as discussed, did not seek to prevent covert content manipulation by the PRC in furtherance of any overarching objective of suppressing (or elevating) certain viewpoints, messages, or content. Supra pp. 14–16. Instead, Congress’s objective was to protect our national security from the clandestine influence operations of a designated foreign adversary, regardless of the possible implications for the mix of views that may appear on the platform.
While that alone sets this case apart from
NetChoice
,
see
2.
The Act’s divestment mandate is narrowly tailored to
achieve Congress’s important national-security interests in
preventing the PRC from accessing U.S. TikTok users’ data
*89
and covertly manipulating content on the platform. The Act
will bring about the severing of PRC control of the TikTok
platform in the United States, either through a divestment of
that control, or, if no qualifying divestment takes place, through
a prohibition on hosting or distributing a still-PRC-controlled
TikTok in the United States until a qualifying divestment
occurs. The divestment mandate is “not substantially broader
than necessary to achieve” Congress’s national-security
objectives.
Ward
,
Congress confined the Act to applications subject to the control of just four designated foreign adversary countries, including China. § 2(g)(4); see 10 U.S.C. § 4872(d)(2). As applied here, the divestment mandate is fashioned to permit the TikTok platform—including its recommendation engine—to continue operating in the United States. Supra p. 10. Insofar as the PRC’s (or ByteDance’s) own decisions may prevent that from happening, the independent decisions of those foreign actors cannot render Congress’s chosen means substantially overbroad.
TikTok submits that various alternate means—including
its proposed National Security Agreement (NSA),
see
Op.,
ante
, at pp. 13–15—would equally fulfill Congress’s aims
without giving rise to the prospect of the platform’s suspended
operations in the United States. But even if we thought that
were true, it would not help TikTok under intermediate
scrutiny: under that standard, “[s]o long as the means chosen
are not substantially broader than necessary,” a law “will not
be invalid simply because a court concludes that the
government’s interest could be adequately served by some less-
speech-restrictive alternative.”
Ward
, 491 U.S. at 800;
see
Turner II
,
Here, Congress reasonably determined that attaining the
requisite degree of protection required mandating a divestment
of PRC control. A “disagreement over the level of
protection . . . to be afforded and how protection is to be
attained” does not constitute a basis for “displac[ing] Congress’
judgment” when applying intermediate scrutiny.
Turner II
,
Nor could TikTok succeed under intermediate scrutiny by pointing to evidence that, in its view, contradicts Congress’s determination that nothing shy of divestment would be sufficient. TikTok argues, for instance, that in concluding the NSA was an inadequate alternative, the government misunderstood certain aspects of its design and operation— e.g., how difficult it would be to review TikTok’s source code. “[R]egardless of whether the evidence is in conflict” on such matters, a court can still sustain a challenged law when applying intermediate scrutiny. Turner II , 520 U.S. at 211. That is because “the relevant inquiry” under that standard is “not whether Congress, as an objective matter, was correct to determine [its chosen means are] necessary” to meet its objectives. Id. ; see id. at 196. “Rather, the question is whether the legislative conclusion was reasonable and supported by substantial evidence in the record before Congress .” Id. at 211 (emphasis added). It was here.
The Executive Branch believed, and specifically advised Congress, that measures short of divestment would not adequately protect against the risks to national security posed by the PRC’s potential control of the TikTok platform. See Newman Decl. ¶ 7 (Gov’t App. 47); Redacted Hearing Tr. 11– *91 14. With specific regard to the provisions contained in the proposed NSA, “senior Executive Branch officials concluded that the terms of ByteDance’s final proposal would not sufficiently ameliorate those risks.” Newman Decl. ¶ 6 (Gov’t App. 46). The provisions, in the Executive Branch’s view, “still permitted certain data of U.S. users to flow to China, still permitted ByteDance executives to exert leadership control and direction over TikTok’s US operations, and still contemplated extensive contacts between the executives responsible for the TikTok U.S. platform and ByteDance leadership overseas.” Id. ¶ 75 (Gov’t App. 62). And, the Executive Branch assessed, the NSA “would have ultimately relied on . . . trusting ByteDance” to comply, but “the requisite trust did not exist.” Id. ¶¶ 75, 86 (Gov’t App. 62, 68).
Those concerns about the kinds of provisions in the NSA
and the overarching lack of trust were discussed with Congress.
See
Redacted Hearing Tr. 10–12, 40–42, 49–50. Congress’s
reliance on those Executive Branch conclusions, even if they
are now disputed by TikTok, means its “legislative conclusion
was . . . supported by substantial evidence in the record before
[it].”
Turner II
,
* * * * *
While the court today decides that the Act’s divestment mandate survives a First Amendment challenge, that is not without regard for the significant interests at stake on all sides. Some 170 million Americans use TikTok to create and view all sorts of free expression and engage with one another and the world. And yet, in part precisely because of the platform’s expansive reach, Congress and multiple Presidents determined that divesting it from the PRC’s control is essential to protect our national security.
To give effect to those competing interests, Congress chose divestment as a means of paring away the PRC’s control—and thus containing the security threat—while maintaining the app and its algorithm for American users. But if no qualifying divestment occurs—including because of the PRC’s or ByteDance’s unwillingness—many Americans may lose access to an outlet for expression, a source of community, and even a means of income.
Congress judged it necessary to assume that risk given the grave national-security threats it perceived. And because the record reflects that Congress’s decision was considered, consistent with longstanding regulatory practice, and devoid of an institutional aim to suppress particular messages or ideas, we are not in a position to set it aside.
Notes
[1] A foreign adversary controlled application is defined in § 2(g)(3) as “a website, desktop application, mobile application, or augmented or immersive technology application that is operated, directly or indi- rectly (including through a parent company, subsidiary, or affiliate), by”: (A) any of — (i) ByteDance, Ltd.; (ii) TikTok; (iii) a subsidiary of or a successor to an entity identified in clause (i) or (ii) that is controlled by a foreign adver- sary; or (iv) an entity owned or controlled, directly or indirectly, by an entity identified in clause (i), (ii), or (iii); or (B) a covered company that — (i) is controlled by a foreign adversary; and (ii) that is determined by the President to present a significant threat to the national security of the United States following [certain procedures].
[2] We use “China” when referring to the country and PRC when referencing its government.
[3] The Executive refers variously to the President, Executive Branch agencies, including the intelligence agencies, and officials thereof.
[4] The term “covered company” is defined as “an entity that operates . . . a website, desktop application, mobile application, or augmented or immersive technology application that”: (i) permits a user to create an account or profile to gener- ate, share, and view text, images, videos, real-time communications, or similar content; (ii) has more than 1,000,000 monthly active users with respect to at least 2 of the 3 months preceding the date on which a relevant determination of the President is made pursuant to paragraph (3)(B); (iii) enables 1 or more users to generate or distribute content that can be viewed by other users of the website, desk- top application, mobile application, or augmented or immersive technology application; and (iv) enables 1 or more users to view content generated by other users of the website, desktop application, mobile
[6] The User Petitioners have not demonstrated that “a substantial
number of” the Act’s “applications are unconstitutional, judged in
relation to the statute’s plainly legitimate sweep.”
Moody v.
NetChoice, LLC
,
[7] Having concluded that TikTok has standing, we need not separately
analyze whether the User Petitioners have standing to raise the same
claims.
See Carpenters Indus. Council v. Zinke
,
[8] We agree with our concurring colleague that the Government’s data-protection rationale “is plainly content-neutral” and standing alone would at most trigger intermediate scrutiny. Concurring Op. 12–13. As we have explained, however, that is not clear for the Government’s content-manipulation justification, and no party has identified any portion of the Act to which the data justification alone applies. We therefore assume strict scrutiny applies to our review of the Act in its entirety and consider both justifications under that standard.
[9] Although our disposition of this case does not turn upon these
reports, the Congress and the President obviously were entitled to
consider such materials when deciding whether to define TikTok as
a foreign adversary controlled application under the Act. Indeed, we
have “approved” the use of similar public materials by the President
when making decisions to designate people or entities under various
national-security related statutes.
See Zevallos v. Obama
, 793 F.3d
106, 109, 113 (2015) (finding it “clear that the government may
decide to designate an entity based on a broad range of evidence,
including intelligence data and hearsay declarations” (quoting
Holy
Land Found. for Relief & Dev. v. Ashcroft
,
[10] The parties offer competing interpretations of this exclusion.
Because we do not doubt the Government’s “proffered . . . interest
actually underlies the law” under either interpretation, we have no
occasion to interpret that provision in this case.
Blount v. SEC
,
[11] Accordingly, we grant the Government’s motion for leave to file classified materials and direct the Clerk to file the lodged materials, though we do not rely on them in denying the petitions.
