Published August 2018
In July 2018 the Internet Commission convened leading UK and international NGOs to discuss emerging approaches to Internet regulation. We posed specific questions to establish key areas of agreement, challenge and disagreement among civil society stakeholders.
These findings will shape the Internet Commission’s work with private firms and governments in the UK and internationally to contribute to a more transparent and accountable Internet. In particular, the Internet Commission will convene industry to develop its thinking on accountability for content management processes. It will also consider the precise role of independent assessment and how the quality of metrics used in transparency reporting could be assured.
Key Challenges | Areas of Debate | Areas of Agreement | |
Policy |
|
|
|
Evidence |
|
|
|
Institutions |
|
|
|
People |
|
|
|
The duty of care proposal could deliver user safety and offer greater clarity about what is illegal. To drive practical improvements, it makes sense to focus on accountability for processes. Key challenges:
Other important debates focus on the implications for intermediary liability and how and when platforms function as public spaces.
Most agree that a strong multi-stakeholder process is required, supported by a range of expert input. It is difficult for politicians to decide on matters of free speech. Governments must be involved, but as one of a range of stakeholders. The key challenge is that clear evidence of harm is often difficult to establish. A risk-based or precautionary approach should therefore be considered.
Strong oversight of processes for removal of illegal and harmful content is needed, including appeals processes. True independence is critical. The multi-stakeholder model can only be effective with the establishment of adequate and sustainable funding for stakeholder representation. Key challenges:
Some argue that there is insufficient technical expertise in policy debates.
There is a wide range of different user needs, but there is also a general lack of digital understanding and so of genuinely informed consent. This is exacerbated by the opacity of the digital advertising ecosystem, which is an obstacle to effective safeguards, transparency, understanding and accountability. Key challenges:
Finally, some argue for a broad definition of platforms to include multiplayer gaming, virtual and augmented reality environments.
In May 2018 the UK Government set out its plan to publish a Code of Conduct and proposals for future legislation [2]. In July 2018, as part of its Dialogue on Digital Responsibility, the Internet Commission brought together leading thinkers on digital responsibility, trust and accountability. They debated the UK Government’s plans and discussed specific proposals from others about transparency reporting [3], an accountability framework [4] and the introduction of a duty of care [5].
Participants included: Age UK, Article 19, Children's Charities' Coalition on Internet Safety, Corsham Institute, defenddigitalme, Index on Censorship, Internet Watch Foundation, NSPCC, Privacy International, UNICEF UK and Which? The Internet Commission is grateful to the Oxford Internet Institute for its support and partnership. The conclusions presented are those of the Internet Commission and do not necessarily reflect the views of all the participants.
Sources:
[1] EUR-Lex Communication on the precautionary principle: http://bit.ly/2LoemdW
[2] UK Government response to its Internet Safety Strategy Green Paper: http://bit.ly/2u1TdvJ
[3] The Internet Commission – transparency reporting framework: http://bit.ly/2KOf2Jw
[4] Keeping Consumers Safe Online – Legislating for platform accountability for online content: http://bit.ly/2zHxlLt
[5] Professor of Internet Law Lorna Woods, University of Essex and William Perrin – written evidence to the House of Lords inquiry, "The Internet: to regulate or not to regulate?": http://bit.ly/2uhdr4t
Want to learn more? We’re happy to help.