引言
隨著人工智能(AI)技術的飛速發展,其對個人數據的處理需求日益增長,這與歐盟《通用數據保護條例》(GDPR)所倡導的嚴格數據保護原則之間形成了微妙的平衡。近年來,歐盟在尋求促進AI創新的同時,也在積極探索如何在現有法律框架下為AI發展提供更大的靈活性。其中,GDPR中的“合法利益”條款(Legitimate Interest)被視為可能為AI訓練“開綠燈”的關鍵路徑。本文將深入剖析歐盟在GDPR修訂和相關指南中對AI訓練合法利益的最新立場,探討其對中國AI企業的潛在影響,並提出相應的合規建議。
核心內容分析
GDPR下“合法利益”條款的概述
GDPR第六條第一款(f)項規定,數據控制者在處理個人數據時,如果為了追求自身或第三方的合法利益,且這種利益不淩駕於數據主體的基本權利和自由之上,則可以合法進行數據處理。這一條款為數據處理提供了除“同意”之外的靈活法律基礎,尤其適用於那些難以獲得數據主體明確同意,但又具有重要商業或社會價值的數據處理活動。
要依賴“合法利益”作為法律基礎,數據控制者必須通過“三步測試”:
-
目的測試(Purpose Test):數據控制者所追求的利益必須是真實、明確且合法的。
-
必要性測試(Necessity Test):數據處理必須是實現該合法利益所必需的,且無法通過其他對數據主體隱私影響更小的方式實現。
-
平衡測試(Balancing Test):數據控制者的合法利益與數據主體的權利和自由之間需要進行權衡,確保數據主體的利益不會被不合理地損害。
歐盟委員會和DPA的最新立場
歐盟的監管機構和數據保護機構(DPA)已就AI訓練中“合法利益”的適用性發布了一系列指南和意見,顯示出在嚴格監管與促進創新之間尋求平衡的趨勢:
-
歐洲數據保護委員會(EDPB)的意見:EDPB在2024年12月發布的意見中明確指出,AI模型的開發和部署可以依賴“合法利益”作為GDPR下的法律基礎。然而,EDPB強調,企業必須進行逐案的合法利益評估(LIA),並實施適當的隱私保障措施,例如數據最小化、假名化以及防止模型輸出中個人數據泄露的措施。
-
法國國家信息與自由委員會(CNIL)的指導:CNIL在2025年6月發布的指導意見中進一步澄清,從公共來源抓取數據用於AI訓練,在滿足特定條件(如尊重網站的robots.txt協議、不從針對未成年人的平臺抓取、不包含高度敏感數據等)下,可以基於合法利益進行。CNIL還指出,即使模型架構使得個體數據的刪除或反對難以直接實現,也可以通過輸出過濾、審計追蹤等替代方案來尊重數據主體權利。
-
“數字綜合法案”提案:2025年11月,歐盟委員會發布了“數字綜合法案”提案,旨在修訂包括GDPR在內的歐盟數字規則。該提案明確提出在GDPR中增設第88c條,規定為開發和運營AI系統而進行的個人數據處理,在符合特定條件且不損害數據主體利益的前提下,可依據合法利益進行。
敏感數據處理的豁免
“數字綜合法案”提案還引入了GDPR第九條的新豁免條款,允許在特定條件和保障措施下,為AI開發和運營處理敏感數據。這解決了AI訓練中大型數據集不可避免地包含少量敏感數據的實際問題。例如,在文本語料庫中包含健康信息,在最小化和保障措施到位的情況下,可能被允許。但需要強調的是,這並非“空白支票”,如果AI系統需要專門處理健康或生物識別數據,仍需獲得明確同意。
匿名化數據定義的拓寬
提案還拓寬了“匿名化”數據的定義,這可能將更多數據處理活動排除在GDPR的適用範圍之外。然而,企業在依賴匿名化數據時,仍需警惕通過再識別攻擊(re-identification attacks)重新識別數據主體的風險。
對中國AI企業的潛在影響和戰略啟示
歐盟的這些最新動態為中國AI企業帶來了機遇,也提出了新的挑戰。
機遇
-
法律確定性增強:歐盟明確將AI訓練納入“合法利益”範疇,為中國企業在歐盟市場開展AI業務提供了更清晰的法律依據,降低了合規不確定性。
-
數據獲取便利:在滿足CNIL等DPA提出的條件和保障措施下,中國企業可以更靈活地利用公共數據進行AI模型訓練,有助於降低數據獲取成本和提高模型效率。
-
敏感數據處理彈性:有限度地處理AI訓練中不可避免的敏感數據,有助於提升AI模型的准確性和魯棒性,尤其是在醫療、金融等對數據質量要求較高的領域。
挑戰
-
合規要求依然嚴格:盡管有所“松綁”,但合法利益評估、數據最小化、保障措施、數據主體權利等GDPR核心要求並未改變,甚至在某些方面有所強化(如反對權被強化為“無條件”)。中國企業需要投入更多資源建立完善的合規體系。
-
跨國監管碎片化:歐盟內部各國DPA對“合法利益”的解釋和執行可能存在差異,加之AI法案、版權法等其他法律框架的疊加,使得合規環境依然複雜。中國企業需要關注不同成員國的具體要求。
-
技術和管理挑戰:實施假名化、匿名化、輸出過濾等技術保障措施,以及建立完善的內部合規流程、進行數據保護影響評估(DPIA)等,對企業的技術和管理能力提出了更高要求。
合規建議
為有效應對歐盟GDPR修訂帶來的機遇與挑戰,中國AI企業應采取以下合規策略:
-
建立健全的合法利益評估機制:嚴格遵循EDPB提出的“目的測試、必要性測試、平衡測試”三步法,對所有基於合法利益的AI數據處理活動進行詳細評估,並妥善記錄評估過程,以備監管機構審查。
-
強化數據保護措施:在AI模型訓練和部署的各個階段,全面實施數據最小化、假名化、匿名化等技術和組織措施。特別是在處理敏感數據時,應采取最先進的安全和隱私保護措施,如嚴格的訪問控制、日志記錄和定期刪除。
-
尊重數據主體權利:確保數據主體的知情權、反對權、刪除權等得到有效保障。即使在模型架構難以直接實現個體數據刪除時,也應提供替代方案,如輸出過濾、審計追蹤或記錄抑制邏輯。
-
關注多重法律框架:除了GDPR,中國AI企業還需密切關注歐盟AI法案、版權法、數字服務法案等相關法規,確保全面合規。例如,在抓取公共數據時,需同時考慮版權和數據庫權利,並尊重平臺的服務條款。
-
持續監測和適應:密切關注歐盟委員會、EDPB和各國DPA發布的最新指南和案例,及時調整內部合規策略和技術實施方案。AI技術的快速發展意味著隱私和合規策略必須保持靈活性和前瞻性 。
結論
歐盟在GDPR修訂和相關指南中對AI訓練“合法利益”的明確,無疑為全球AI產業,包括中國AI企業,帶來了積極信號。這表明歐盟在平衡數據保護與技術創新方面邁出了重要一步。然而,這種“松綁”並非監管的放松,而是對合規路徑的進一步明確。中國AI企業應抓住這一機遇,積極調整合規策略,建立健全的數據保護體系,以專業的洞察力和前瞻性的法律商業建議,在歐盟市場實現可持續發展。
Prospects of GDPR Revision (II): Green Light for AI Training? How the “Legitimate Interest” Clause Empowers Chinese AI Companies
Introduction
With the rapid development of artificial intelligence (AI) technology, the demand for processing personal data has been increasing, creating a delicate balance between such demand and the stringent data protection principles advocated by the European Union’s General Data Protection Regulation (GDPR). In recent years, the EU has been actively exploring ways to facilitate AI innovation while providing greater flexibility for AI development within the existing legal framework. Among these, the “Legitimate Interest” clause under the GDPR is seen as a key pathway to “greenlight” AI training. This article will provide an in-depth analysis of the EU’s latest stance on the legitimate interest basis for AI training in the GDPR revision and related guidelines, explore its potential impact on Chinese AI enterprises, and propose corresponding compliance recommendations.
What is a Data Protection Core Content AnalysisOfficer (DPO)?
Overview of the “Legitimate Interest” Clause under GDPR
Article 6(1)(f) of the GDPR stipulates that a data controller may lawfully process personal data if it is necessary for the purposes of the legitimate interests pursued by the controller or by a third party, provided such interests are not overridden by the fundamental rights and freedoms of the data subject. This clause offers a flexible legal basis for data processing beyond “consent,” especially for data processing activities for which it is difficult to obtain explicit consent for but hold significant commercial or social value.
To rely on legitimate interest as a legal basis, data controllers must undergo a “three-step test”:
-
Purpose Test: The pursued interest must be genuine, explicit, and lawful.
-
Necessity Test: The data processing must be necessary to achieve that legitimate interest and cannot be realized through other means that would have less impact on the data subject’s privacy.
-
Balancing Test: A balance must be struck between the data controller’s legitimate interests and the rights and freedoms of the data subjects, ensuring that the latter’s interests are not unreasonably harmed.
Latest Positions from the European Commission and DPAs
EU regulators and Data Protection Authorities (DPAs) have issued a series of guidelines and opinions regarding the applicability of legitimate interest in AI training, demonstrating a trend of balancing strict regulation with fostering innovation:
-
European Data Protection Board (EDPB) Opinion: In the opinion released in December 2024, the EDPB clearly stated that AI model development and deployment can rely on legitimate interest as a legal basis under the GDPR. However, it emphasized that companies must conduct case-by-case legitimate interest assessments (LIAs) and implement appropriate privacy safeguards, such as data minimization, pseudonymization, and measures to prevent leakage of personal data in model outputs.
-
French National Commission on Informatics and Liberty (CNIL) Guidance: CNIL’s guidance issued in June 2025 further clarifies that data scraping from public sources for AI training can be based on legitimate interest under specific conditions—such as respecting the website’s robots.txt protocol, avoiding platforms targeting minors, and excluding highly sensitive data. CNIL also noted that even when the model architecture makes direct deletion or objection of individual data difficult, alternative solutions like output filtering and audit trails can be employed to respect data subject rights.
-
“Digital Comprehensive Act” Proposal: In November 2025, the European Commission proposed the “Digital Comprehensive Act,” aiming to revise EU digital regulations including the GDPR. The proposal explicitly introduces Article 88c to the GDPR, stipulating that personal data processing for the development and operation of AI systems may be based on legitimate interest under specific conditions and without harming data subject interests.
Exemptions for Processing Sensitive Data
The “Digital Comprehensive Act” proposal also introduces new exemptions under Article 9 of the GDPR, allowing the processing of sensitive data for AI development and operation under certain conditions and safeguards. This addresses the practical issue that large AI training datasets inevitably contain some sensitive data. For example, health information included in text corpora may be permitted if minimization and protective measures are in place. It is important to stress, however, that this does not constitute a “blank check”; explicit consent remains required when AI systems specifically process health or biometric data.
Broadening the Definition of Anonymized Data
The proposal further broadens the definition of “anonymized” data, potentially excluding more data processing activities from GDPR’s scope. Nonetheless, companies relying on anonymized data must remain vigilant against re-identification risks through re-identification attacks.
Potential Impact and Strategic Insights for Chinese AI Companies
These latest EU developments present both opportunities and challenges for Chinese AI enterprises.
Opportunities
-
Enhanced Legal Certainty: The EU’s explicit inclusion of AI training under legitimate interest provides Chinese companies with a clearer legal basis for conducting AI business in the EU market, reducing compliance uncertainties.
-
Facilitated Data Access: Under the conditions and safeguards proposed by CNIL and other DPAs, Chinese companies can more flexibly use public data for AI model training, helping reduce data acquisition costs and improve model efficiency.
-
Flexibility in Sensitive Data Processing: The limited allowance for processing sensitive data unavoidable in AI training helps improve model accuracy and robustness, especially in data-quality demanding fields like healthcare and finance.
Challenges
-
Strict Compliance Requirements Remain: Despite some relaxation, core GDPR requirements such as legitimate interest assessments, data minimization, safeguards, and data subject rights remain unchanged or are even strengthened (e.g., the right to object is reinforced as “unconditional”). Chinese enterprises need to invest more resources to develop comprehensive compliance systems.
-
Fragmented Cross-border Regulation: Variations in interpretation and enforcement of legitimate interest among EU member states’ DPAs, combined with overlapping legal frameworks such as the AI Act and copyright law, continue to complicate the compliance environment. Chinese companies must pay close attention to specific requirements in different member states.
-
Technical and Managerial Challenges: Implementing pseudonymization, anonymization, output filtering, and establishing robust internal compliance processes including Data Protection Impact Assessments (DPIAs) pose higher demands on enterprises’ technical and management capabilities.
Compliance Recommendations
To effectively respond to the opportunities and challenges brought by the GDPR revision, Chinese AI companies should adopt the following compliance strategies:
-
Establish Robust Legitimate Interest Assessment Mechanisms: Strictly follow the EDPB’s three-step approach—purpose test, necessity test, and balancing test—for all AI data processing activities based on legitimate interest. Document the assessment process thoroughly to prepare for regulatory inspections .
-
Enhance Data Protection Measures: Implement comprehensive technical and organizational safeguards such as data minimization, pseudonymization, and anonymization throughout all stages of AI model training and deployment. When processing sensitive data, apply state-of-the-art security and privacy protections including strict access controls, logging, and regular data deletion.
-
Respect Data Subject Rights: Ensure effective guarantees of data subjects’ rights to information, objection, and deletion. Even when direct deletion is difficult due to model architecture, provide alternative solutions like output filtering, audit trails, or suppression logic records.
-
Monitor Multiple Legal Frameworks: Beyond GDPR compliance, Chinese AI companies must also closely follow relevant EU regulations such as the AI Act, copyright law, and Digital Services Act to ensure comprehensive compliance. For example, when scraping public data, also consider copyright and database rights and respect platform terms of service.
-
Maintain Continuous Monitoring and Adaptation: Keep abreast of the latest guidelines and case law issued by the European Commission, EDPB, and national DPAs, and timely adjust internal compliance strategies and technical implementations. The fast evolution of AI technology requires privacy and compliance strategies to remain flexible and forward-looking.
Conclusion
The EU’s clarification of legitimate interest for AI training in the GDPR revision and related guidelines undoubtedly sends positive signals to the global AI industry, including Chinese AI enterprises. It demonstrates an important step forward in balancing data protection with technological innovation. However, this “loosening” does not imply regulatory relaxation but rather a clearer path for compliance. Chinese AI companies should seize this opportunity to actively adjust compliance strategies and establish sound data protection systems, leveraging professional insights and forward-looking legal and business advice to achieve sustainable development in the EU market.
聲明
本文僅為交流探討之目的,不代表廣悅律師事務所或其律師出具的任何形式之法律意見或建議。如需轉載或引用本文的任何內容,請與本所溝通授權事宜,並於轉載或引用時注明出處。如您有意就相關業務進一步交流或探討,或需要專業的法律支持,歡迎與本所聯系。


聯系人:葉文女士
期待與您的進一步交流!
廣悅律師事務所介紹
廣悅律師事務所成立於2008年,是一家立足大灣區,堅持一體化管理的涉外綜合性律師事務所。發展至今,廣悅建立了由上百位律師及其他法律服務人員組成的專業團隊,打造了多元化的業務體系,可以為客戶提供高品質、全方位、一站式的法律服務。秉承“立足灣區、協同港澳、面向世界”的發展戰略,廣悅已擁有廣州、中國香港、深圳,以及泰國曼穀、美國洛杉磯、澳大利亞悉尼、日本東京、意大利米蘭八個辦公室,客戶遍及境內外多個國家和地區。
供稿丨廣悅米蘭辦公室
編輯丨吳寶渲
審核丨蘇 冰
審定丨品牌宣傳與市場拓展委


