[Dcpr] Due Diligence Recommendations

Julia Powles jep50 at cam.ac.uk
Thu Jan 8 00:33:24 EST 2015


Hi Luca and DC PR members,

I had a few comments on this document I thought I'd send on, taking your 
suggestion to circulate to the mailing-list. It concerns an important 
subject to which you've clearly dedicated considerable work and 
generated a commendable draft. Hopefully these thoughts are of some use 
-- but if they are too wayward, that's ok too.

All best- Julia

Some thoughts:

1)	Since a specific objective is concision (“the spectrum of rights and 
remedies that are granted to platform users through the ToS may be 
difficult to comprehend or even read in its entirety”, also III.A of the 
recommendations) and harmonisation (“and similar platforms may be 
regulated through very different provisions that might be unilaterally 
modified by platform providers”), my impression is that the document is 
both too long, and too inspecific. Length may be necessary, but in this 
case, it does not fully deliver on the laudable promise of offering 
practical red-lines beyond the old OECD rules and five mandatory 
conditions (my point 2 below), nor provide examples about what is meant 
by, for example, specific consent and specific purposes, and what would 
definitely *not* be ok, or definitely *would* be ok. It seems to me that 
somewhat bright lines are what is needed for easy semantic comparison of 
ToS and for clear implementation by programmers and guarantees by 
operators.

2)	There are, as far as I can tell, only five mandatory conditions - is 
this right?:
(a) specific rather than general consent for data collection by category 
(II.A, II.D), involving real choice with no coercion (Annex 1 (j));
(b) right to opt out of behavioural tracking – what does this mean? 
Tracking is not defined. And, arguably, it extends to most data 
collection, so this to me seems incompatible with the rest of this 
section (II.A);
(c) right to find out information about predictive or probabilistic 
techniques used and their rationale – really? Algorithms are famously 
guarded… what do you mean here? (II.D);
(d) right to permanently delete account (II.D);
(e) prohibition on platform operator unilaterally terminating account 
without appropriate grounds (III.A).
Small point: given that III is prefaced as discussing non-derogable 
procedures (i.e. mandatory “shall” conditions), do you want to be more 
specific about what they actually are in this context? Or otherwise 
eliminate that language, as it’s confusing.
Also, it seems to me that the protection of children should have some 
more mandatory conditions, not least because informed consent has an age 
threshold.

3)	There should be examples – what would be appropriate for a specified 
purpose? Can a platform say: “we collect personal information to make 
our service more relevant and personalised to you”? What does specific 
consent mean? II.D goes some way towards addressing this, but it could 
be clearer, and better integrated with the rest of the document. How 
does this document address and set an example about short, concise ToS 
that do something short of making you sign over your first child to very 
long-winded ToS because there’s no practical alternative?

4)	Consistency is lacking between many sections. E.g., if specific 
consent is mandatory in II.A & II.D, then it is inconsistent that 
automatically scanning for advertising in para 2 of II.A. is not also 
mandatory, or that II.C or II.E is not – how can use beyond original 
purpose be specifically consented to, or actual usage by third parties? 
And is the implication that specific consent does not extend to 
knowledge and consent to data retention (II.B), as opposed to 
collection? Further it seems inconsistent that “everyone’s right to 
remove incorrect or excessive personal data” is mentioned at the end of 
V, but not at all earlier in the document, at least not as a mandatory 
right in such terms.

5)	Part of what you are trying to address is global operations, but you 
don’t offer protections for users from any state’s meddling, apart from 
the transparency reporting (II.E). Is “legitimate law” the law of the 
data subject, the operator, the operations?! Is there a way of excluding 
every single possible country that might have an interest in a user's 
data?

6)	The absence of mention of copyright under IV is noticeable, even 
though I think it appropriate for it not to be discussed on the same 
level. Nevertheless, it is the most significant factor in takedowns on 
online platforms.

Anyway, as I say, these are just some thoughts that I hope are helpful 
and which I think would further your objectives, although I recognise 
you are fairly well progressed in the project. Best of luck and hope to 
see you in the year ahead.







> ---------- Forwarded message ----------
> From: <LB at lucabelli.net>
> Date: Fri, Dec 26, 2014 at 3:31 PM
> Subject: [Dcpr] Due Diligence Recommendations
> To: dcpr at lists.platformresponsibility.info
> 
> Dear all,
> 
> Thanks a lot for your very constructive inputs and critiques with
> regard to the original draft of the Due Diligence Recommendations. We
> have significantly amended the original draft compiling your comments
> and suggestions. You may find the revised version both in attachment
> and below (the version below does not have footnotes).
> 
> By all means, this is not the final version yet and you are all
> encouraged to share your comments on this second draft so that we can
> furhter enhance it. Ideally, the final version of the Due Diligence
> Recommendations should be released on 31 JANUARY.
> 
> The second comment period is now open and will last until 20 JANUARY.
> Although we have greatly appreciated your comments via personal
> emails, we kindly suggest that you circulate your future comments on
> the mailing-list to trigger discussion amongst the DC PR members.
> Alternatively, you may also comment the Due Diligence Recommendations
> using this google doc
> https://docs.google.com/document/d/1N0_aVbWSNt-S12O0KLebqYVxM6gj5Uz3qqHDUkmlPc0/edit?pli=1
> [1]#
> 
> Lastly, you are all encouraged to share any relevant info (articles,
> posts, events, etc.) related to online platforms’ responsability to
> respect human rights, using this mailing-list.
> 
> As this year is ending we wish you, your families and friends an
> excellent year ahead.
> All the best,
> 
> Luca, Nicolo and Primavera
> 
> DRAFT
> 
> Dynamic Coalition on Platform Responsibility
> 
> -------------------------
> 
> Due Diligence Recommendations
> on Terms of Service and Human Rights
> 
> -------------------------
> 
> Introduction
> 
> The following recommendations aim at fostering online platforms’
> responsibility to respect human rights, by providing guidance with
> regard to the adoption of “responsible” terms of service. Besides
> identifying minimum standards for the respect of human rights by
> platform operators (standards that “shall” be met), these
> recommendations  suggest best practices (which “should” be
> followed) for the most “responsible” adherence to human rights
> principles in the drafting of terms of service.
> 
> 	*
> Background
> 
> Aside from cases of complicity in human rights violations rising to
> the level of international crimes over which the International
> Criminal Court has jurisdiction, private entities cannot be held
> liable under international law for human rights violations, as they
> are no parties to human rights treaties. Yet, respect of human rights
> undoubtedly represents an important factor to take into account when
> assessing the activities of corporations from the perspective of a
> variety of stakeholders, including governments, investors and
> increasingly, consumers.
> 
> This is especially relevant in the context of online platforms
> designed to serve the needs of a global community, and forced to
> satisfy different, often conflicting legal requirements across the
> various jurisdictions where they operate. Thus, in light of the
> important role that online platforms are playing in shaping a global
> information society and the significant impact they have on the
> exercise of the rights of Internet users, a moral and social
> expectation has formed that such entities behave “responsibly”,
> ensuring the respect of the rule of law and the human rights of the
> Internet users across the globe.
> 
> The existence of a responsibility of private sector actors to respect
> human rights, which was recently affirmed in the UN Guiding Principles
> on Business and Human Rights and unanimously endorsed by the UN Human
> Rights Council, is grounded upon the tripartite framework developed by
> the UN Special Rapporteur for Business and Human Rights, according to
> which States are the primary duty bearers in securing the protection
> of human rights, corporations have the responsibility to ensure the
> respect of human rights, and both entities are joint duty holders in
> providing effective remedies against human rights violations.
> 
> As part of this responsibility, corporations are expected to: (1) make
> a policy commitment to the respect of human rights; (2) adopt a human
> rights due-diligence process to identify, prevent, mitigate and
> account for how they address their impacts on human rights; and (3)
> have in place processes to enable the remediation of any adverse human
> rights impacts they cause or to which they contribute.
> 
> These recommendations focus on one of the most concrete and tangible
> means for online platforms to bring that responsibility to bear: the
> contractual agreement which Internet users are required to adhere to
> in order to utilise their services (the so called “Terms of
> Service”), thus becoming platform users. Specifically, the
> recommendations constitute an attempt to define “due diligence”
> standards for online platforms with regard to four essential
> components: privacy, freedom of expression, due process and protection
> of children and young people. In doing so, they aim to provide a
> benchmark for respect of human rights in the private governance
> sphere, both in the relation of a platform’s own conduct as well as
> with regard to the scrutiny of governmental requests that they
> receive.  As recently stressed by the Council of Europe’s
> Commissioner for Human Rights, guidance on these matters is
> particularly important due to the current lack of clear standards.
> This applies a fortiori in the context of online platforms, given the
> crucial role that these entities have in ensuring practical compliance
> with fundamental rights on the Internet.
> 
> 	*
> Privacy
> 
> The first section of these recommendations provides guidance over the
> rules that online platform operators should adopt in order to ensure
> the protection of their users against any unnecessary and unreasonable
> collection, use and disclosure of personal data.
> 
> 	*
> Data Collection
> 
> Platform operators should limit the collection of personal information
> to what is directly relevant and necessary to accomplish a specified
> purpose. They shall also obtain consent for every type of information
> collected (by category), rather than through a single general-purpose
> consent form. If consent is withdrawn, the platform is no longer
> entitled to process such data. Although withdrawal is not retroactive,
> i.e. it cannot invalidate the data processing that took place in the
> period during which the data was collected legitimately, it should, in
> principle, prevent any further processing of the individual’s data
> by the controller.
> 
> In principle, platform operators should also refrain from collecting
> data by automatically scanning content privately shared by their
> users. Admissible derogations to this principle include the need to
> fight against unsolicited communications (spam) and to ensure network
> security, and should not extend to commercial or advertising purposes
> in the absence of specific and express platform-user consent.
> 
> Platform operators shall always offer their users the possibility to
> opt out from the tracking of their behavior both by the platform
> within other services, and by other services within the platform.
> 
> In order to facilitate user oversight on the application of these
> principles, platform operators should  allow their users to view, copy
> and modify the personal information they have made available to the
> platform, and are encouraged to do so enabling download of a copy of
> their personal data in interoperable format.
> 
> 	*
> Data Retention
> 
> Platform operators should clearly communicate in their terms of
> service whether and for how long they are storing any personal data.
> As a general rule, any retention beyond 180 days should be
> specifically justified by a function of the platform or by
> requirements imposed by a  “legitimate law”.
> 
> 	*
> Data aggregation
> 
> Aggregation of platform users’ data should only be done subject to
> express consent. Aggregation of data across multiple services or
> devices requires extra diligence from the part of the data controller
> and processor, since it might result in data being processed beyond
> the original purpose for which it was collected. Although this does
> not prevent the implementation of cross-device functionalities, it is
> necessary to ensure that platform users properly understand the scope
> of the given consent.
> 
> 	*
> Data Use
> 
> Platforms shall obtain consent in order to use personal data
> (including platform users’ contacts and recipients), unless such use
> is prescribed by a legitimate law.  It is also recommended that such
> consent is required for platforms to make personal data available to
> the public through search engines, and to give their  users the option
> to opt out.
> 
> The requirement of consent also apply to personal data over which the
> platform acquires the right to use the data therein; as a general
> rule, such use should never be broader than the original purpose for
> which personal data was shared. Oftentimes, the use of personal data
> is instrumental to the improvement of existing services, or the
> development of new services and functionalities. Yet, a broad and
> open-ended permission on the use of platform users’ personal data
> for “future services” can lead to abuses, in particular by making
> it possible for the platform to offer personalised services on the
> basis of the information provided, and automatically enroll them into
> services or functionalities that they did not intend to receive at the
> time of registration. This is in conflict with the right of users to
> informational self-determination, including the right not to be
> subject to any decision based solely on automated processing of data
> or without taking their view into consideration. For this reason, it
> is recommended that platforms specify in their ToS that the purpose of
> processing of personal data is limited to the scope of existing
> services, and the enrolment of platform users into any new service
> will require their acceptance to new ToS. Platform operators should
> also give them the opportunity to object to such usage and  demand the
> rectification of inaccurate data. Furthermore, platform users shall
> always be able to obtain information about any predictive or
> probabilistic techniques they have been used to build their profile,
> and their underlying rationale.
> 
> Lastly, platform operators shall always permit their users to delete
> their account in a permanent fashion. Likewise, if there is no other
> legal reason justifying the further storage of the data, the data
> processor shall proceed with the permanent deletion of all or portions
> of the relevant data associated with the platform user’s account, in
> a time that is reasonable for its technical implementation.
> 
> 	*
> Data protection vis-à-vis third parties
> 
> Platforms play a crucial role in enforcing the protection offered by
> the legal system against interference with users’ right to privacy.
> Therefore, platform operators should provide effective remedies
> against the violation of internationally recognised human rights. For
> this reason, they should establish clear mechanisms for platform users
> to gain access to all of their personal data held by a third party, as
> well as to be informed of the actual usage thereof. Platform operators
> should also enable their users to report privacy-offending content and
> to submit takedown requests.They should also implement a system to
> prevent the impersonation of platform users by third parties, although
> exceptions would need to be made for the impersonation of public
> figures in ways which contributes to the public debate in a democratic
> society.
> 
> A second set of concerns pertains to the possibility to preempt any
> interference with platform users’ personal data, by preventing third
> parties’ access to platform user’s content and metadata. Firstly,
> platform operators should allow users to preserve their anonymity
> vis-à-vis third parties to the extent permitted by legitimate laws,
> both within the platform itself and within other websites when such
> platform is used as an identity service. Secondly, it is recommended
> that platforms enable end-to-end encryption of communications and
> other personal information, in the context of both storage and
> transmission. In that respect, best practice is when the decryption
> key is retained by the platform user, except where the provider needs
> to hold the decryption key in order to provide the service and the
> platform user has provided informed consent.
> 
> As regards the handing over of platform users’ data upon
> governmental request, platform operators should specify that they
> execute such request only in the presence of a valid form of legal
> process, and release a periodic transparency report providing, per
> each jurisdiction in which they operate, the amount and type of such
> requests, and the platforms’ response (in aggregate numbers).
> 
> 	*
>  Due Process
> 
> Due process is a fundamental requirement for any legal system based on
> the rule of law. “Due” process refers to the non-derogability of
> certain procedures in situations which may adversely affect
> individuals within the legal system. This includes the application of
> minimum safeguards such as the clarity and predictability of the
> substantive law, the right to an effective remedy against any human
> rights violation and the right to be heard before any potentially
> adverse decision is taken regarding themselves.
> Due process has significant implications with regards to potential
> amendment and termination of contractual agreements, as well as the
> adjudication of potential disputes. This section aims to give content
> the due diligence in this context.
> 
> 	*
> Amendment and termination of contracts
> 
> Terms of Service should be written in plain language that is easy to
> understand. Wherever possible, the platform operators should provide
> an accessible summary of the key provisions of the terms of service.
> The platform operators should give its users meaningful notice of any
> amendment of the ToS affecting the rights and obligation of the users.
> Meaningful notice should be provided in a way that enables platform
> users to clearly see, process and understand the changes. Contractual
> clauses that permit unilateral termination by platforms without
> appropriate grounds shall not be used.
> 
> In addition, platform operators should consider giving notice even of
> less significant changes, and enabling their users to access previous
> versions of the terms of service. Ideally, platforms operators should
> enable their users to continue using the platforms without having to
> accept the new terms of service related to the additional
> functionalities which have been added to the platform, unless this
> would impose significant costs and complexity to the operators.
> Meaningful notice should also be given in advance, prior to
> termination of the contract or services . Besides, to reduce the
> imbalance between platform users and platforms owners when it comes to
> litigation, it is recommendable that the ToS be negotiated beforehand
> with consumer associations or other organisations representing
> Internet users. In order to prevent wrongful decisions, it is also
> recommended that platforms make termination of accounts of particular
> platform users possible only upon repeated violation of ToS.
> 
> 	*
> Adjudication
> 
> Disputes can arise both between platform users and between a
> particular platform user and the platform operator. In both cases,
> platform operators should provide alternative dispute resolutions
> systems to allow for quicker and potentially more granular solutions
> than litigation for the settling of disputes. However, in view of the
> fundamental importance of the right of access to court, alternative
> dispute resolution systems should not be presented as a replacement of
> regular court proceedings, but only in addition to those. In
> particular, platform operators should not impose waiver of class
> action rights or other hindrances to the right of an effective access
> to justice, such as mandatory jurisdiction outside the place of
> residence of Internet users. Any dispute settlement mechanism should
> be clearly explained and offer the possibility of appealing against
> the final decision.
> 
> 	*
> Freedom of Expression
> 
> Freedom of expression is a fundamental right consisting of the freedom
> to receive and impart information in a lawful manner, including by way
> of association. In the online platform context, the effectiveness of
> this right can be seriously undermined by disproportionate monitoring
> of online speech and repeated government blocking and takedown. The
> following section provides guidance as to how platforms should handle
> such matters through their terms of service.
> 
> 	*
> Degree of monitoring
> 
> Although there are no rules to determine, in general terms, what kind
> of speech should or should not be allowed in private online platforms,
> certain platforms should be seen more as “public spaces” to the
> extent that occupy an important role in the public sphere. These
> actors have assumed functions in the production and distribution
> process of media services which, until recently, had been performed
> only (or mostly) by traditional media organisations. As a matter of
> fact, online platforms increasingly play an essential role of speech
> enablers and pathfinders to information, becoming instrumental for
> media’s outreach as well as for Internet users’ access to them.
> 
> Established principles and best-practices can serve to identify
> certain red-lines that should not be crossed. As a general rule,
> platform operators should not impose any restrictions on the kind of
> content that they host, with the exception of content that is harmful
> or explicitly forbidden under applicable legitimate laws (e.g. hate
> speech, child pornography and incitement to violence, as well as other
> kinds of undesirable content, such as unsolicited communications for
> direct marketing purposes or security threats) but only if necessary
> and proportionate to that purpose. It is of utmost importance that the
> rules imposing such restrictions are not formulated in such a way as
> to affect potentially legitimate content, as they would otherwise
> constitute a basis for censorship.
> 
> Similarly, although platforms can legitimately remove speech in order
> to enforce their terms of service, either on their own motion or upon
> complaint, such terms of service should be clear and transparent in
> their definition of the content that will be restricted within the
> platform, including the use of certain screen names. To this end,
> platforms can legitimately prohibit the use of the name, trademark or
> likeness of others. In addition, platforms operator should always
> provide clear mechanisms to notify those platform users whose content
> has been removed and provide them with an opportunity to challenge and
> override illegitimate restrictions.
> 
> 	*
> Government blocking and takedowns
> 
> Transparent procedures should be adopted for the handling and
> reporting of governmental requests for blocking and takedown in a way
> that is consistent with internationally recognised laws and standards.
> Firstly, platform operators should execute such requests only where
> there is a legal basis for doing so and a valid judicial order.
> Secondly, platforms operators should notify their users of such
> requests, ideally giving them an opportunity to reply and challenge
> the validity of such requests, unless specifically prohibited by
> legitimate law. Finally, as already mentioned in the context of
> government requests for data, platform operators should implement law
> enforcement guidelines and release periodic transparency reports.
> 
> V. Protection of children and young people
> 
> A special category of concerns arises in the case of children and
> young people, towards which platform operators should exercise a
> higher level of care. Platform operators should adopt particular
> arrangements, beyond the mere warning for inappropriate content and
> age verification that can be imposed by legitimate law for certain
> types of content.
> 
> Firstly, although terms of service should generally be drafted in a
> way that is comprehensible to all, those regulating platforms open to
> children and young people should include facilitated language or an
> educational video-clip and, ideally, a set of standardised badges  to
> make their basic rules comprehensible by all users regardless of their
> age and willingness to read the actual terms of use. Secondly, it is
> recommended that platforms provide measures that can be taken by
> children and young people in order to protect themselves while using
> the platform, such as utilising a “safer navigation” mode.
> Thirdly, platform operators should consider providing a specific
> mechanisms to ensure removal or erasure of content created by children
> and young people.
> As an element of media literacy, all platform users should be informed
> about their right to remove incorrect or excessive personal data.
> 
> Annex 1: Definitions
> 
> a) Platform:
> For the purpose of these recommendations, platforms are understood as
> cloud-based public-facing interfaces or “spaces” allowing users to
> impart and receive information or ideas according to the rules defined
> into a contractual agreement.
> 
> b) Terms of Service:
> The concept of “terms of service” utilised here covers not only
> the contractual document available under the traditional heading of
> “terms of service” or “terms of use”, but also any other
> platform’s policy document (e.g. privacy policy, community
> guidelines, etc.) that is linked or referred to therein.
> 
> c) Function of the Platform:
> Function that the community has attributed to the platform on the
> basis of the legal, commercial and social expectations that it has
> generated. This should not be confused with a platform’s
> functionalities, which constitute merely one (albeit important)
> element to identify the overall function(s).
> 
> d) Platform Operator
> Natural or legal person defining and having the possibility to amend
> the platform’s terms of service
> e) Platform User
> Natural or legal person entering into a contractual relationship
> defined by the platform’s terms of service.
> f) Internet User
> Natural or legal person who is using Internet access service, and in
> that capacity has the freedom to impart and receive information. The
> Internet user may be the subscriber, or any person to whom the
> subscriber has granted the right to use the Internet access service
> s/he receives.
> g) Data:
> Content and/or personal information. Data can belong to both
> categories simultaneously.
> 
> h) Content:
> Text, image, audio or video which results from the engagement of a
> particular platform user with the platform, even on a transient basis.
> This includes, for example, messages and queries typed by a platform
> user.
> 
> i) Personal Data/Personal Information:
> Personal data is any information about an individual that can be used
> to distinguish or trace an individual’s identity, such as name,
> social security number, date and place of birth, etc. This is not
> intended to cover identification which can be accomplished via very
> sophisticated methods. This notion of personal data equates with that
> of Personally Identifiable Information (PII), defined as “any
> information about an individual maintained by an agency, including (1)
> any information that can be used to distinguish or trace an
> individual‘s identity, such as name, social security number, date
> and place of birth, mother‘s maiden name, or biometric records; and
> (2) any other information that is linked or linkable to an individual,
> such as medical, educational, financial, and employment
> information.”
> j) Consent:
> Consent means any freely given, specific and informed indication of
> the data subject’s wishes by which s/he signifies  her/his agreement
> to personal data relating to her/himself being processed. To that end,
> every user shall be able to exercise a real choice with no risk of
> deception, intimidation, coercion or significant negative consequences
>  if he/she does not consent to data aggregation.
> 
> k) Express Consent:
> Express consent is a type of consent which (in contrast with
> “implicit” or “implied” consent) requires an affirmative step
> in addition to the acceptance of the general ToS, such as clicking or
> ticking a specific box or acceptance of the terms and conditions of a
> separate document.
> 
> l) Privacy:
> Privacy is an inalienable human right enshrined in Article 12 of the
> Universal Declaration of Human Rights, which establishes the right of
> everyone to be protected against arbitrary interference with their
> privacy, family, home or correspondence, and against attacks upon his
> honour and reputation. In the context of online platforms, this
> encompasses the ability for data subjects to determine the extent to
> which and the purpose for which their personal data is used by data
> controllers, including the conditions upon which such data can be made
> available to third parties (right to informational
> self-determination).
> 
> m) Freedom of Expression:
> The right to freedom of expression, enshrined in article 19 of the UN
> Declaration of Human Rights, includes freedom to hold opinions without
> interference and to seek, receive and impart information and ideas
> through any media and regardless of frontiers. The right to freedom of
> opinion and expression is as much a fundamental right on its own
> accord as it is an “enabler” of other rights, including economic,
> social and cultural rights.
> 
> n) Hate Speech:
> Although there is no universally accepted definition of “hate
> speech”, the term shall be understood as covering all forms of
> expression which spread, incite, promote or justify racial hatred,
> xenophobia, anti- Semitism or other forms of hatred based on
> intolerance, including: intolerance expressed by aggressive
> nationalism and ethnocentrism, discrimination on any grounds such as
> race, ethnicity, colour, sex, language, religion, political or other
> opinion, national or social origin, property, disability, birth,
> sexual orientation or other status. In this sense, “hate speech”
> covers comments which are necessarily directed against a person or a
> particular group of persons.
> 
> o) Due Process:
> Due process is a concept referring to procedural rights which are
> essential for the respect of the rule of law, comprising: (1) the
> right to an effective remedy by a competent tribunal for any acts
> violating one’s fundamental rights granted by the law or the
> Constitution (ex article 8 of the Universal Declaration of Human
> Rights); and (2) and the right to an independent and impartial
> tribunal, in the determination of one’s rights and obligations and
> of any criminal charge against him/her (ex. art.10 of the Universal
> Declaration  of Human Rights). In the context of online platforms, the
> urgency and efficacy of the protection of these rights might require
> an expansion of the scope of application of these rights beyond the
> traditional notion of “tribunal”.
> 
> p) Legitimate Law:
> Laws and regulations should be deemed as legitimate when they respond
> to a pressing social need and, having regard to their tangible impact,
> they can be considered as proportional to the aim pursued.
> The concept of “legitimate law” which is used to justify a
> potential restriction shall be interpreted as referring to a law or
> regulation which does not manifestly fail the requirements for
> permissible restrictions  identified for freedom of expression (and
> applicable, mutatis mutandis, to restrictions of other fundamental
> rights) by the UN Special Rapporteur on the promotion and protection
> of the right to freedom of expression and opinion, specifically:
> (a) It must be provided by law, which is clear and accessible to
> everyone (principles of predictability and transparency);
> (b) It must pursue a legitimate purpose (principle of legitimacy); and
> (c) It must be proven as necessary and the least restrictive means
> required to achieve the purported aim (principles of necessity and
> proportionality).
> If it is manifest that the measure would not pass these three-pronged
> test, the platform operator should deny the request and, to the extent
> possible, challenge it before the relevant court.
> 
> _______________________________________________
>  DCPR mailing list
>  DCPR at lists.platformresponsibility.info
>  http://lists.platformresponsibility.info/listinfo/dcpr [2]
> 
> 
> 
> Links:
> ------
> [1]
> https://docs.google.com/document/d/1N0_aVbWSNt-S12O0KLebqYVxM6gj5Uz3qqHDUkmlPc0/edit?pli=1
> [2] http://lists.platformresponsibility.info/listinfo/dcpr


More information about the DCPR mailing list