This file was downloaded from://eprints.qut.edu.au/127442/1/Submission+on+AI... · 2020. 12....

46
This may be the author’s version of a work that was submitted/accepted for publication in the following source: Guihot, Michael & Rimmer, Matthew (2019) Artificial Intelligence: Governance and Leadership - A submission to the Australian Human Rights Commission and World Economic Forum. Australian Human Rights Commission and World Economic Forum, Aus- tralia. This file was downloaded from: https://eprints.qut.edu.au/127442/ c 2019 the Author(s) This work is covered by copyright. Unless the document is being made available under a Creative Commons Licence, you must assume that re-use is limited to personal use and that permission from the copyright owner must be obtained for all other uses. If the docu- ment is available under a Creative Commons License (or other specified license) then refer to the Licence for details of permitted re-use. It is a condition of access that users recog- nise and abide by the legal requirements associated with these rights. If you believe that this work infringes copyright please provide details by email to [email protected] Notice: Please note that this document may not be the Version of Record (i.e. published version) of the work. Author manuscript versions (as Sub- mitted for peer review or as Accepted for publication after peer review) can be identified by an absence of publisher branding and/or typeset appear- ance. If there is any doubt, please refer to the published source. https:// tech.humanrights.gov.au/ consultation

Transcript of This file was downloaded from://eprints.qut.edu.au/127442/1/Submission+on+AI... · 2020. 12....

  • This may be the author’s version of a work that was submitted/acceptedfor publication in the following source:

    Guihot, Michael & Rimmer, Matthew(2019)Artificial Intelligence: Governance and Leadership - A submission to the

    Australian Human Rights Commission and World Economic Forum.Australian Human Rights Commission and World Economic Forum, Aus-tralia.

    This file was downloaded from: https://eprints.qut.edu.au/127442/

    c© 2019 the Author(s)

    This work is covered by copyright. Unless the document is being made available under aCreative Commons Licence, you must assume that re-use is limited to personal use andthat permission from the copyright owner must be obtained for all other uses. If the docu-ment is available under a Creative Commons License (or other specified license) then referto the Licence for details of permitted re-use. It is a condition of access that users recog-nise and abide by the legal requirements associated with these rights. If you believe thatthis work infringes copyright please provide details by email to [email protected]

    Notice: Please note that this document may not be the Version of Record(i.e. published version) of the work. Author manuscript versions (as Sub-mitted for peer review or as Accepted for publication after peer review) canbe identified by an absence of publisher branding and/or typeset appear-ance. If there is any doubt, please refer to the published source.

    https:// tech.humanrights.gov.au/ consultation

    https://eprints.qut.edu.au/view/person/Guihot,_Michael.htmlhttps://eprints.qut.edu.au/view/person/Rimmer,_Matthew.htmlhttps://eprints.qut.edu.au/127442/https://tech.humanrights.gov.au/consultation

  • AUSTRALIAN HUMAN RIGHTS COMMISSION

    AND WORLD ECONOMIC FORUM

    ARTIFICIAL INTELLIGENCE: GOVERNANCE AND LEADERSHIP

    QUT Robotronica 2015

    DR MICHAEL GUIHOT

    SENIOR LECTURER

    QUT COMMERCIAL AND PROPERTY LAW CENTRE

    FACULTY OF LAW

    QUEENSLAND UNIVERSITY OF TECHNOLOGY

    DR MATTHEW RIMMER

    PROFESSOR OF INTELLECTUAL PROPERTY AND INNOVATION LAW

    FACULTY OF LAW

    QUEENSLAND UNIVERSITY OF TECHNOLOGY

    Queensland University of Technology

    2 George Street GPO Box 2434

    Brisbane Queensland 4001 Australia

    Work Telephone Number: (07) 31381599

  • 2  

    EXECUTIVE SUMMARY

    This submission addresses the white paper on Artificial Intelligence: Governance and

    Leadership produced by the Australian Human Rights Commission. In particular, it focuses

    upon the dimensions of intellectual property, commercial law, and regulation.

    In regards to intellectual property, there has been significant case law, particularly in

    the United States dealing with the interaction of established IP laws with developments

    in AI and robots.

    Commercial aspects of the development of AI and robots should be governed using

    either the Competition and Consumer Act 2010 and the Australian Consumer Law or

    modifications to them to specifically address problems.

    There are a number of regulatory responses that can respond to developments in AI and

    robots including legislative amendments, self-regulation and soft law approaches such

    as nudging. The significant commercial impact of developments in these new

    technologies will require a hardened and practiced regulator such as the ACCC to be

    effective.

    Recommendation 1

    Intellectual property law plays a key role in the regulation of artificial intelligence, and other

    related fields of technology. Intellectual property holders will hold considerable influence in

    terms of the use and exploitation of artificial intelligence technologies. Australia has a diverse

    array of regulators in the field. IP Australia has oversight of industrial forms of property – such

    as patents, trade marks, and designs. The Department of Communications and the Arts has

    carriage of copyright law. The Australian Competition and Consumer Commission also plays

    a role in relation to misleading or deceptive conduct. At an international level, the World

    Intellectual Property Organization has played a significant role in tracking technology trends

    in respect of AI, particularly through patent information. There may well need to be reforms to

    intellectual property law, policy, and practice in light of developments in AI.

  • 3  

    Recommendation 2

    The Australian Competition and Consumer Commission, or a newly created technology

    subdivision of the ACCC should be the body that overseas and enforces the amended

    legislation in relation to consumer transactions that involve problems associated with

    developments in AI. This is for two reasons:

    Firstly, the ACCC already has expertise in developing, enforcing and educating

    Australians on regulation built up over 20 years of experience in consumer protection.

    Secondly, there is already a vast and powerful regulatory enforcement regime in place

    under the Competition and Consumer Act that could, if need be, be amended to apply

    to problems associated with developments in AI.

    Recommendation 3

    If a new body such as the proposed Responsible Innovation Organisation is created, its role

    should be limited to education and coordination between the various regulatory bodies

    regulating AI. Because the rate and degree of change in AI development is so rapid and deep,

    and the possible uses to which it might be put is unknowable, no single agency could maintain

    full vigilance or control over these developments. If that proves to be the case, then any single

    agency that did take on a governance role would likely fail. Any agency that does take on a

    governance role and fails, will consequently bear some liability if its governance is lacking, for

    whatever reason. Setting up an agency that is not able to fulfil its role would merely transfer

    some, if not all, of the liability for problems caused by AI to the agency, away from the

    technology companies. We must be careful not to shift the burden from the

    manufacturer/supplier to authority.

  • 4  

    Definition and Classification

    Lin, Abney and Bekey define a robot as:

    … an engineered machine that senses, thinks, and acts: “Thus a robot must have sensors, processing

    ability that emulates some aspects of cognition, and actuators. … on-board intelligence is necessary if

    the robot is to perform significant tasks autonomously…”1

    Thus AI and robots lie together on a spectrum, with AI as the thinking part of the ‘sense, think,

    act’ paradigm. Calo differentiates between AI and robots such that ‘a technology does not act,

    and hence is not a robot, merely by providing information in an intelligible format. It must be

    in some way.’2 The embodiment or physicality required of a robot takes AI into the world.

    Balkin too is open to the idea that AI and robots are merely elements on a spectrum:

    as Calo points out, there is a continuum between “robots” and “artificial intelligence.” That is because,

    like the Internet itself, robots and other interactive entities do not have to be designed in any particular

    way. And because there is a continuum of potential designs and a variety of different potential uses, there

    is also a continuum of potential effects that these new technologies can have.3

                                                                1 Patrick Lin, Keith Abney and George Bekey, ‘Robot Ethics: Mapping the Issues for a Mechanized

    World’ (2011) 175(5–6) Artificial Intelligence 942, 943 citing; George A Bekey, Autonomous Robots: From

    Biological Inspiration to Implementation and Control (MIT Press, 2005). 2 Ryan Calo, ‘Robotics and the Lessons of Cyberlaw’ (2015) 103(3) California Law Review 513, 531. 3 Jack M Balkin, ‘The Path of Robotics Law’ (2015) 6 California Law Review Circuit 45, 50.

  • 5  

    In the last twenty years or so, there has been an exponential increase in the power, speed, and

    capacity of computers. At the same time, the stockpile of machine-accessible data has similarly

    multiplied. The machine learning capabilities of Artificial intelligence (AI) now has the

    dexterity to traverse these data troves, using algorithms to make conclusions based on

    connections that humans are incapable of seeing. Companies use AI for hiring decisions,4 to

    profile customers, and to present customised information on social media sites. Courts and

    judges use AI in sentencing, bail and probation decisions. As computers become more

    pervasive, we continue to accede to and trust their ‘superior’ decision-making abilities.

    However, there is a growing concern about the use of, and outcomes achieved by, some

    automated decision-making processes. These concerns include concerns about incursions into

    consumers’ informational and personal privacy. Also, the complex and opaque applications

    that use decision-making algorithms are often ‘black boxes’5 and, as such, the decisions they

    make often cannot be examined or explained easily. Further, the data used to train the models

    is sometimes, itself, flawed, incomplete, or may even entrench existing biases.6 This relentless

    advancement of automation and AI in the private sector has roused public discourse about the

    need for regulatory oversight.7 Given the potential impact of AI applications, researchers,

    journalists, data scientists, lawyers and policy makers have an obligation to mitigate threats or

    risks caused by AI, including by questioning and testing every step in the development of new

    technologies.

                                                                4 Simon Chandler, The AI Chatbot Will Hire You Now (13 September 2017)

    https://www.wired.com/story/the-ai-chatbot-will-hire-you-now/. 5 Frank Pasquale, The Black Box Society: The Secret Algorithms That Control Money and Information

    (Harvard Univ. Press, 2015). 6 Solon Barocas and Andrew D Selbst, ‘Big Data’s Disparate Impact’ (2016) 104 California Law Review

    671. 7 Oren Etzioni, ‘Opinion | How to Regulate Artificial Intelligence’ The New York Times, 20 January 2018

    https://www.nytimes.com/2017/09/01/opinion/artificial-intelligence-regulations-rules.html; The United Nations

    Department of Economic & Social Affairs, ‘The Impact of the Technological Revolution on Labour Markets and

    Income Distribution’ (The United Nations, 31 July 2017) https://www.un.org/development/desa/dpad/wp-

    content/uploads/sites/45/publication/2017_Aug_Frontier-Issues-1.pdf; Michael Guihot, Anne Matthew and

    Nicolas Pierre Suzor, ‘Nudging Robots: Innovative Solutions to Regulate Artificial Intelligence’ (2017) 20(2)

    Vanderbilt Journal of Entertainment and Technology Law 385.

  • 6  

    The work of Roger Brownsword suggests that there are recurring themes in how the

    legal system tries to deal with new, disruptive technologies.8 In the fields of law and

    technology, there are recurring debates over how laws should be technology-neutral,

    technology-contextual, or technology-specific. There has certainly been such a debate over

    robotics law and policy, and discussion as to whether it is an autonomous field, or lacks a

    specific identity.

    There has also been a great discussion about liability rules in respect of robotics, and

    significant debate over legal rules regarding automation in transportation. Both automobile

    manufacturers and information technology companies have been engaged in research and

    development over autonomous vehicles. There has been significant debate over the road rules

    for autonomous vehicles – such as Google’s self-driving car. Likewise, drones have raised

    challenging policy questions in respect of aviation rules. The appearance of aquabots has also

    posed intriguing matters about the law of sea. The adoption of robotics in agriculture has also

    raised questions about automation. In the field of health care, the use of robotics holds out the

    promise of improving health outcomes for patients. Yet, given the past conflicts over medical

    liability, there is a need to lay down appropriate rules, standards, and codes about the use of

    robotics in the areas of surgery, patient care, and prosthetics.

    As well as the discussion about civilian uses of robots, there has also been much interest

    in the increasing use of robots by law enforcement agencies. At an international level, there

    has been deep disquiet about the use of drone warfare by major superpowers. There has been a

    movement to ban ‘killer robots’.

    There has been significant debate about the impact of robots, automation, and artificial

    intelligence upon employment. Optimists hope that the robotics revolution will result in the

    creation of new jobs. Pessimists fear that automation will lead to redundancies, under-

    employment, and underemployment across a range of industries. One policy recommendation

    has been that there should be a robot tax to generate funds for training of workers, in areas such

    as manufacturing, who are displaced by automation.

                                                                8 Roger Brownsword, Eloise Scotford, and Karen Yeung, Oxford Handbook in Law, Regulation and

    Technology. Oxford: Oxford University Press, 2015.

  • 7  

    Bill Gates has been enthusiastic about the idea of taxing robotics.9 However, critics

    have complained that special forms of taxation in respect of robotics would discourage

    research, development, and innovation.

    Cory Doctorow has wondered whether it is even possible to regulate robots. He argues:

    ‘I am skeptical that "robot law" can be effectively separated from software law in general.’10

    Doctorow, channelling Lawrence Lessig, suggests that we think about how software and

    robotics is regulated ‘through code, through law, through markets and through norms.’11

    Mark Lemley and Bryan Casey have discussed the debate over the definition and

    classification of robots and artificial intelligence.12 They have highlighted problems of over-

    inclusive and under-inclusive definitions of robots and artificial intelligence. Lemley and

    Casey argue:

    Rather than trying in vain to find the perfect definition, we instead argue that policymakers should do

    as the great computer scientist, Alan Turing, did when confronted with the challenge of defining robots:

    embrace their ineffable nature. We offer several strategies to do so. First, whenever possible, laws

    should regulate behavior, not things (or as we put it, regulate verbs, not nouns). Second, where we must

    distinguish robots from other entities, the law should apply what we call Turing’s Razor, identifying

    robots on a case-by-case basis. Third, we offer six functional criteria for making these types of “I know

    it when I see it” determinations and argue that courts are generally better positioned than legislators to

    apply such standards. Finally, we argue that if we must have definitions rather than apply standards,

    they should be as short-term and contingent as possible. That, in turn, suggests regulators—not

    legislators—should play the defining role.13

    This seems to be a plea for a pragmatic approach to the definition, classification, and regulation

    of robotics and artificial intelligence.

                                                                9 Kevin Delaney, ‘The Robot That Takes Your Job Should Pay Taxes, Says Bill Gates’, Quartz, 18

    February 2017, https://qz.com/911968/bill-gates-the-robot-that-takes-your-job-should-pay-taxes/ 10 Cory Doctorow, ‘Why It Is Not Possible to Regulate Robots’, The Guardian, 2 April 2014,

    https://www.theguardian.com/technology/blog/2014/apr/02/why-it-is-not-possible-to-regulate-robots 11 Ibid. 12 Mark Lemley and Bryan Casey, ‘You Might Be a Robot’, SSRN, 2019,

    https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3327602#.XHsk_VR5I_Q.twitter 13   Ibid. 

  • 8  

    1. Intellectual Property

    Francis Gurry - the director-general of the World Intellectual Property Organization – has

    discussed the need for research in respect of intellectual property and artificial intelligence:

    AI is fast becoming part of our everyday lives, changing how we work, shop, travel and interact with

    each other. Yet we are only at the beginning of discovering the many ways in which AI will have an

    impact on – and indeed challenge – business, society and culture. There are numerous misconceptions

    and misgivings about the nature of AI, and in particular the challenge it poses to humankind. Given

    these widely held reservations and concerns, it is essential to have a factual basis for policy discussions

    about innovation in AI.14

    He observes that intellectual property plays a key role in the regulation of artificial intelligence

    and robotics. In particular, the disciplines of trade secrets, patents, designs, trade marks, and

    copyright law will impinge upon

    A. Trade Secrets, Confidential Information, and Privacy

    New Innovations in robotics and artificial intelligence have often been protected through trade

    secrets, and confidential information. Keisner, Raffo and Wunsch-Vincent comment that ‘the

    technological complexity of robotics systems means that trade secrets are often the first option

    for companies seeking to protect their innovations’.15

    There have been other disputes over trade secrets relating to the field of robotics

    (broadly construed). Notably, in 2018, there was a conflict between Waymo and Uber relating

    to confidential information and patents associated with autonomous vehicles (self-driving

    cars). After going to trial, the matter was ultimately settled in favour of Waymo. The dispute

    is nonetheless an important precedent in respect of intellectual property, employment law, and

    robotics.

    In May 2017, a Seattle jury decided that Huawei misappropriated trade

    secrets belonging to T-Mobile related to a smartphone-testing robot called Tappy, which T-

                                                                14 WIPO, Technology Trends 2019: Artificial Intelligence, Geneva: WIPO, 2019, 7. 15 C. Andrew Keisner, Julio Raffo and Sacha Wunsch-Vincent, ‘Breakthrough Technologies – Robotics

    and IP’, WIPO Magazine, 2016.

  • 9  

    Mobile had in its Bellevue laboratory.16 The jury determined that Huawei had misappropriated

    T-Mobile's trade secrets in a series of incidents in 2012 and 2013, and that it breached a handset

    supply contract between the two companies. The jury said T-Mobile should be awarded $4.8

    million in damages because of this breach of contract. In January 2018, the United States

    Department of Justice brought a criminal action against Huawei for the theft of trade secrets

    held by T-Mobile in respect of the robot Tappy.17 The dispute in USA v. Huawei will be a major

    precedent on intellectual property and robotics.18

    The current trade dispute between the United States and China is amongst other things

    about intellectual property, technology transfer, and innovation policy. The Made in China

    2025 innovation policy is particularly focused upon China developing capacity in fields of

    advanced manufacturing – including robotics. The United States is keen to preserve its

    competitive advantage in industry and technology.

    In October 2017, U.S. prosecutors dropped charges against Dong Liu - a dual citizen of

    China and Canada - accused of trying to steal trade secrets from Medrobotics - a Massachusetts-

    based manufacturer of robotic surgical products - by trespassing at its headquarters (Raymond,

    2017). There seemed to be a lack of evidence in the end for the matter to proceed.

    In 2018, an automotive robotics supplier exposed documents detailing assembly line

    schematics, robotics configurations, and other trade secrets of Toyota, Tesla, and Volkswagen

    on a publicly accessible server. This case highlighted issues about information security in the

    robotics sector.

    In 2018, the AI Now Institute led by Kate Crawford has argued: ‘AI companies should

    waive trade secrecy and other legal claims that stand in the way of accountability in the public

    sector’.19 The group has elaborated:

    Vendors and developers who create AI and automated decision systems for use in government should

    agree to waive any trade secrecy or other legal claim that inhibits full auditing and understanding of

                                                                16 Rachel Lerman, ‘Jury Awards T-Mobile $4.8m in Trade-Secrets Case against Huawei’, The Seattle

    Times, 20 May 2017. 17 United States Department of Justice, ‘Chinese Telecommunications Device Manufacturer and its U.S.

    Affiliate Indicted for Theft of Trade Secrets, Wire Fraud, and Obstruction Of Justice,’ Press Release, 28 January

    2019; and John Marsh, ‘Tappy's Revenge: What You Need to Know About the DOJ's Momentous Trade Secret

    Indictment of Huawei’, Lexblog, 7 February 2019. 18 USA v. Huawei (2019), Case 2:19-cr-0010-RSM 19 AI Now Institute, AI Now Report, December 2018, https://ainowinstitute.org/AI_Now_2018_Report.pdf

  • 10  

    their software. Corporate secrecy laws are a barrier to due process: they contribute to the “black box

    effect” rendering systems opaque and unaccountable, making it hard to assess bias, contest decisions,

    or remedy errors. Anyone procuring these technologies for use in the public sector should demand that

    vendors waive these claims before entering into any agreements.20

    The group was keen to ensure that the auditability and transparency of AI systems was not

    unduly affected by trade secrets and confidential information.

    Moreover, the AI Now Institute argue: ‘Technology companies should provide

    protections for conscientious objectors, employee organizing, and ethical whistleblowers.’

    They comment:

    Organizing and resistance by technology workers has emerged as a force for accountability and ethical

    decision making. Technology companies need to protect workers’ ability to organize, whistleblow, and

    make ethical choices about what projects they work on. This should include clear policies

    accommodating and protecting conscientious objectors, ensuring workers the right to know what they

    are working on, and the ability to abstain from such work without retaliation or retribution. Workers

    raising ethical concerns must also be protected, as should whistleblowing in the public interest.21

    In this context, there is a need to ensure that there are proper defences and exceptions for

    whistleblowers and conscientious objectors under trade secrets law.

    In response to such intellectual property claims in the technological field, some have

    instead looked to open licensing in respect of robotics.

    As well as raising matters of trade secret protection and confidential information,

    artificial intelligence and robotics – and similar technologies like autonomous vehicles and

    drones – raise issues about privacy. Calo, Froomkin and Kerr have highlighted the significant

    privacy implications of robotics in their respective work.22 The Australian Parliamentary

    inquiry into autonomous vehicles was particularly concerned about data collection and data

    gathering by self-driving cars, and the privacy ramifications of such activities. The Queensland

    Government is currently holding an inquiry into the privacy implications of surveillance by

    drones.

                                                                20 Ibid. 21 Ibid. 22 Ryan Calo, A. Michael Froomkin, and Ian Kerr (ed.), Robot Law, Cheltenham and Northampton (MA):

    Edward Elgar, 2016.

  • 11  

    B. Patent Law

    The United States Patent and Trademark Office (USPTO) has developed a classification for

    robots. This collection ‘provides for a reprogrammable, multifunction manipulator designed to

    move devices through variable programmed motions for the performance of changeable tasks

    on a repetitive basis without human intervention and all subcombinations thereof specialized

    for use with such manipulator.’ The USPTO relies upon a definition of an industrial robot from

    the International Organization for Standardization.

    In the field of patent law, there has been significant patent activity in respect of robotics.

    The WIPO report on breakthrough innovation charts the geography of patent activity in the

    area of robotics.23 Japanese, Korean, and German companies dominated the top rankings for

    filing patents in the area of robotics. China was notably improving in its performance

    (particularly in terms of public sector patent filings). There is a need to improve Australia’s

    performance in translating research into practical outcomes. The 2019 WIPO study has

    highlighted the patterns of patent activity in respect of artificial intelligence.24 While the

    classifications for robotics and artificial intelligence are distinct, there is a significant

    intersection and overlap between the two fields of technology.

    There has been increasing litigation in respect of patents relating to robots and robotics.

    In 2011, a United States District Court found that all Genmark Automation patent claims at

    issue in a legal action were valid and infringed by Innovative Robotics Systems Inc, and entered

    a final consent judgment and permanent injunction against any further infringement of

    Genmark’s patents.

    In 2017, iRobot Corp, a leader in consumer robots and maker of the Roomba®

    vacuuming robot, filed legal proceedings for patent infringement against multiple robotic

    vacuum cleaner manufacturers and sellers including Bissell, Hoover, bObsweep, iLife, Black

    & Decker and the Chinese or Taiwanese companies that manufacture the infringing products.

    In 2017, the US International Trade Commission investigated whether iRobot’s patents in

    respect of robotics had been infringed by rival robots sold by Bissell, Hoover, Black & Decker

    and others. In 2018, the International Trade Commission made a final determination, which

    bars products from Hoover and bObsweep from importation into the United States. The action

                                                                23 WIPO, World IP Report: Breakthrough Innovation and Economic Growth, Geneva: WIPO, 2015. 24 WIPO, Technology Trends 2019: Artificial Intelligence, Geneva: WIPO, 2019.

  • 12  

    has led to other cases against iLife, Bissell, and Black & Decker, among others, to end in

    favourable settlements to iRobot.

    There has been concern about non-practising patent entities filing patent infringement

    actions in the field of robotics (much as there has been issues in the field of information

    technology). One suit against Intuitive Surgical was brought by Alisanos LLC after Medicanica

    – a company that retained surgical robotics patents – transferred its patent portfolio to

    Alisanos.25

    In 2019, Chelmsford's Endeavor Robotics – which designed a ‘Scorpion Robot for the

    U.S. Army - sued defense firm QinetiQ North America for patent infringement. Endeavor

    Robotics claims that QinetiQ infringed on two of its patents, one for a ‘robotic platform’ and

    one for a ‘mobile robotic vehicle’. Endeavor claims that QinetiQ North America’s robot for

    the U.S. Army’s Common Robotic System-Individual program infringed its stair-climbing

    robots.

    WIPO commented: ‘One can start to see the more intensive offensive and defensive IP

    strategies that are present in other high-technology fields.’26 WIPO has wondered: ‘A vital

    question is whether the increased stakes and commercial opportunity across various sectors

    will tilt the balance toward costly litigation, as in other hightech and complex technologies.’27

    It will also be worthwhile seeing whether patent defences and exemptions are deployed in

    respect of robotics – particularly in respect of the defence of experimental use; compulsory

    licensing; and crown use. There may also be scope for patent pools and public sector licensing

    in respect of robotics.

    As part of its ‘Made in China 2025’ strategy, China has developed its own local robotics

    industry to boost advanced manufacturing.28 In order to compete with its foreign competitors,

    China has established a patent pool in respect of robotics technology. The patent pool is

    intended to be a means of tackling obstacles to the development of China’s robotics industry,

    such as a lack of core patents.

    Other possible regulatory tools include funding research and development – such as

    innovation prizes and challenges in the field of robotics.

                                                                25 Tim Sparapani, ‘Surgical Robotics and the Attack of Patent Trolls’, Forbes, 19 June 2015. 26 WIPO, World IP Report: Breakthrough Innovation and Economic Growth, Geneva: WIPO, 2015, 129. 27 Ibid., 129. 28 CHOFN, ‘Country Developing New Age of Robotics with Patent Efforts’, 7 March 2016.

  • 13  

    In its 2019 study, WIPO has considered the intellectual property trends in respect of

    artificial intelligence.29 In its executive summary, WIPO highlights the wide variety of

    applications of artificial intelligence:

    Artificial intelligence (AI) is increasingly driving important developments in technology and business,

    from autonomous vehicles to medical diagnosis to advanced manufacturing. As AI moves from the

    theoretical realm to the global marketplace, its growth is fueled by a profusion of digitized data and

    rapidly advancing computational processing power, with potentially revolutionary effect: detecting

    patterns among billions of seemingly unrelated data points, AI can improve weather forecasting, boost

    crop yields, enhance detection of cancer, predict an epidemic and improve industrial productivity.30

    WIPO also noted that some technologies had multiple applications: ‘Many AI-related

    technologies can find use across different industries, as shown by the large number of patents

    in AI that refer to multiple industries.’31

    Analysing the patent data, WIPO highlights the dominance of entities from United

    States, China, and Japan. The report observes:

    Companies, in particular those from Japan, the United States of America (U.S.) and China, dominate

    patenting activity. Companies represent 26 out of the top 30 AI patent applicants, while only four are

    universities or public research organizations. This pattern applies across most AI techniques,

    applications and fields. Of the top 20 companies filing AI-related patents, 12 are based in Japan, three

    are from the U.S. and two are from China. Japanese consumer electronics companies are particularly

    heavily represented.32

    In this context, Australia is in a precarious position – lagging in many of the key innovation

    races in respect of patents and AI.

    In terms of patent applicants, WIPO charted the dominance of IBM and Microsoft:

    IBM and Microsoft are leaders in AI patenting across different AI-related areas IBM has the largest

    portfolio of AI patent applications with 8,290 inventions, followed by Microsoft with 5,930. Both

    companies’ portfolios span a range of AI techniques, applications and fields, indicating that these

    companies are not limiting their activity to a specific industry or field. Rounding out the top five

                                                                29 WIPO, Technology Trends 2019: Artificial Intelligence, Geneva: WIPO, 2019. 30 Ibid., 13. 31 Ibid., 14. 32 Ibid., 15.

  • 14  

    applicants are Toshiba (5,223), Samsung (5,102) and NEC (4,406). The State Grid Corporation of

    China has leaped into the top 20, increasing its patent filings by an average of 70 percent annually from

    2013 to 2016, particularly in the machine learning techniques of bio-inspired approaches, which draw

    from observations of nature, and support vector machines, a form of supervised learning.33

    The heavy dominance of AI patents by corporations will also have important implications in

    terms of the ownership of AI, and access to benefits associated with AI.

    The report also discusses the activity of universities and public research institutions in

    the context of AI research. The report comments: ‘Despite the dominance of companies in AI,

    universities and public research organizations play a leading role in inventions in selected AI

    fields such as distributed AI, some machine learning techniques and neuroscience/

    neurorobotics.’34 The report highlights the significant investment by Chinese universities in AI

    patents:

    Chinese organizations make up 17 of the top 20 academic players in AI patenting as well as 10 of the

    top 20 in AI-related scientific publications. Chinese organizations are particularly strong in the

    emerging technique of deep learning. The leading public research organization applicant is the Chinese

    Academy of Sciences (CAS), with over 2,500 patent families and over 20,000 scientific papers

    published on AI. Moreover, CAS has the largest deep learning portfolio (235 patent families). Chinese

    organizations are consolidating their lead, with patent filings having grown on average by more than

    20 percent per year from 2013 to 2016, matching or beating the growth rates of organizations from

    most other countries.35

    The report observes that there are 167 universities and public research organisations ranked

    among the top 500 patent applicants: ‘110 are Chinese, 20 are from the U.S., 19 from the

    Republic of Korea and 4 from Japan; [and] four European public research organizations feature

    in the top 500 list.’36

    C. Designs Law

                                                                33 Ibid., 15. 34 Ibid., 16. 35 Ibid., 16. 36 Ibid., 15.

  • 15  

    Designs law may also raise issues in respect of the design of robots and robotics. Keisner, Raffo

    and Wunsch-Vincent observe that ‘industrial designs that protect a robot’s appearance – its

    shape and form – also play an important role in improving the marketability of products and

    helping firms appropriate the returns on their R&D investments.’37

    As highlighted by the litigation between Apple and Samsung over designs relating to

    smartphones and tablets, designs law can play a significant role in disputes over the ownership

    of new technologies.

    The right of repair under designs law will also play an important role in respect of

    robotics. The Federal Court of Australia has provided guidance as to the nature and scope of

    the right of repair in GM Global Technology Operations LLC v. S.S.S. Auto Parts Pty Ltd.38

    There has been increasing activity in respect of algorithm-driven design. There has been

    a conversation about artificial intelligence will impact the work of designers.

    D. Trademark Law and Related Rights

    In addition to other fields of industrial property law, there have also been battles over trade

    marks and robotics. The makers of the film RoboCop have asserted their trademark against

    providers of security services. Lucasfilm – which developed the Star Wars franchise – acquired

    a trademark on ‘Droid.’ Trade marks will play a critical role as robotics companies seek to

    market their inventions in the global marketplace.

    There has been consideration as to how artificial intelligence will affect trade mark law.

    Lee Curtis and Rachel Platts of HGF explain how the AI revolution will impact upon the legal

    discipline.39 They explain:

    The impact of AI systems in everyday life and the process of buying products and services, which in

    essence is the focus of trademark law, is increasing. It is predicted by a study from Gartner that by

    2020, 85% of customer service interactions in retail will be powered or influenced by some form of AI

    technology. AI global revenue is predicted by market intelligence firm Tractica to skyrocket from

                                                                37 C. Andrew Keisner, Julio Raffo and Sacha Wunsch-Vincent, ‘Breakthrough Technologies – Robotics

    and IP’, WIPO Magazine, 2016. 38 GM Global Technology Operations LLC v. S.S.S. Auto Parts Pty Ltd [2019] FCA 97 39 Lee Curtis and Rachel Platts, ‘AI Is Coming and It Will Change Trade Mark Law’,

    http://www.hgf.com/media/1173564/09-13-AI.PDF

  • 16  

    $643.7 million in 2016 to $36.8 billion in 2025. A report from advertising agency J WalterThompson

    suggests that 70% of so-called millennials appreciate brands using AI technology to showcase their

    products, with a report from Statista suggesting that 38% percent of consumers receive better

    purchasing guidance with AI than without.40

    Curtis and Platts note: ‘To date, AI and IP discussions have centred around patent law and

    patent protection for AI software applications.’41 However, they observed: ‘The impact of AI

    on trade mark law and whether the present law is “fit for purpose” seems to have been

    completely overlooked.’42

    There has been one piece of litigation, which involved artificial intelligence and trade

    mark law. In Cosmetic Warriors and Lush v Amazon.co.uk and Amazon EU, Lush argued that

    Amazon had infringed its trade marks.43 There was a consideration of Amazon’s use of product

    suggestion:

    In connection with the search engine on its own site, Amazon has also used analyses of consumer

    behaviour. Thus for example, if a consumer types Squiffo into the search box and that term has not

    been typed in previously, no results will be shown and the screen may ask if the consumer meant

    “squiff” and display some results for squiff products. However, if the consumer who originally typed

    in squiffo goes on to purchase some goods, these goods might be suggested to the next consumer who

    types in squiffo. It is for reasons like this that consumers who type Lush into the amazon.co.uk search

    facility are shown products such as Bomb Cosmetics products—previous consumers who typed in Lush

    have gone on to browse and/or purchase such products. Thus, Amazon has built up and uses a

    behaviour-based search tool to identify an association between a particular search word and specific

    products. Amazon uses this tool to present products to consumers which it hopes will be of interest to

    them. In the present case, this tool has used the word Lush to identify products which Amazon believes

    a consumer searching for Lush products might wish to buy instead of a Lush-branded product.44

    Baldwin J held that those sponsored advertisements for Amazon, triggered by keywords

    including ‘Lush’, which included the mark, infringed Article 5(1)(a) of the Directive. However

    those advertisements that did not include the LUSH mark were not infringing.

                                                                40 Ibid. 41 Ibid. 42 Ibid. 43 Cosmetic Warriors and Lush v Amazon.co.uk and Amazon EU [2014] EWHC 181 (Ch). 44 Cosmetic Warriors and Lush v Amazon.co.uk and Amazon EU [2014] EWHC 181 (Ch).

  • 17  

    There were also some disputes over publicity rights and robotic representations in the

    1990s in the United States in a number of cases such as Wendt v. Host International, Inc., and

    White v. Samsung Electronics.45 There has been discussion of whether there is a need to revisit

    legal personhood in light of robotics and artificial intelligence.

    The advent of robotics and artificial intelligence may also have implications for

    consumer law and competition policy. Kate Darling has emphasized the need to take into

    account the public interest in competition policy in matters of intellectual property and robotics:

    ‘Competition will drive better implementations of personalized robots, and a vibrant market

    could even encourage better privacy and data security solutions.’46

    E. Copyright Law

    Copyright law and technological protection measures are relevant to robotics through the

    means of protecting computer programs, databases, and creative works.

    In copyright law, robotics poses complicated questions about authorship, ownership,

    and creativity. At the QUT Robotronica festivals, there have been a number of demonstrations

    of how robotics has been transforming the creative arts – including in music, art, and

    performance.

    In the private sector, information technology companies have experimented with

    machine learning, neural networks, and artificial intelligence.

    There has been significant debate as to whether the copyright categories of authorship

    could include artificial intelligence. In the IceTV v. Nine Network, the High Court of Australia

    has insisted upon the need for human authorship of copyright works.47 The judges stressed the

    importance of human authorship and human agency:

    The first principle concerns the significance of "authorship". The subject matter of the Act now extends

    well beyond the traditional categories of original works of authorship, but the essential source of

    original works remains the activities of authors. While, by assignment or by other operation of law, a

    party other than the author may be owner of the copyright from time to time, original works emanate

                                                                45 Wendt v. Host International, Inc., 125 F.3d 806 (1997) and White v. Samsung Electronics 971 F.2d 1395

    (1992) 46 Kate Darling, ‘Why Google’s Robot Personality Patent Is Not Good for Robotics’, IEEE Spectrum, 8

    April 2015. 47 IceTV v. Nine Network [2009] HCA 14.

  • 18  

    from authors. So it was that in Victoria Park Racing and Recreation Grounds Co Ltd v Taylor, Dixon

    J observed:

    "Perhaps from the facts a presumption arises that the plaintiff company is the owner of the

    copyright but, as corporations must enlist human agencies to compose literary, dramatic,

    musical and artistic works, it cannot found its title on authorship. No proof was offered that the

    author or authors was or were in the employment of the company under a contract of service

    and that the book was compiled or written in the course of such employment."

    Key provisions of Pt III of the Act fix on "the author". Examples include the requirement for the author

    of unpublished works to be a "qualified person" for copyright to subsist (s 32(1)), the fixing of copyright

    duration by reference to the death of the author (s 33), and the conferral of copyright upon the author

    subject to the terms of employment or contractual arrangements under which the author labours (s 35).48

    The High Court of Australia emphasized that ‘the notion of "creation" conveys the earlier

    understanding of an "author" as "the person who brings the copyright work into existence in its

    material form"’. 49

    In the United States, a number of jurists and legal theorists have considered ways and

    means by which robotics and artificial intelligence could be accommodated within copyright

    law. Pamela Samuelson,50 Annemarie Bridy,51 James Grimmelmann,52 and Andres Guadamuz53

    have explored the possibilities in respect of copyright law and artificial intelligence.

    Presciently, in 1985, Pamela Samuelson from Berkeley Law considered the question of

    allocation of ownership rights in computer-generated works. 54 She observed: ‘As “artificial

    intelligence” (AI) programs become increasingly sophisticated in their role as the “assistants”

    of humans in the creation of a wide range of products – from music to architectural plans to

                                                                48 IceTV v. Nine Network [2009] HCA 14. 49 IceTV v. Nine Network [2009] HCA 14. 50 Pamela Samuelson, ‘Allocating Ownership Rights in Computer-Generated Works’ (1985) 47 University

    of Pittsburgh Law Review 1185-1228. 51 Annemarie Bridy, ‘Coding Creativity: Copyright and the Artificially Intelligent Author’ (2012) 5

    Stanford Technology Law Review 1-28. 52 James Grimmelmann, ‘Copyright for Literate Robots’ (2016) 101 Iowa Law Review 657-681. 53 Andres Guadamuz, ‘Artificial Intelligence and Copyright’, WIPO Magazine, October 2017.

    https://www.wipo.int/wipo_magazine/en/2017/05/article_0003.html 54 Pamela Samuelson, ‘Allocating Ownership Rights in Computer-Generated Works’ (1985) 47 University

    of Pittsburgh Law Review 1185-1228.

  • 19  

    computer chip designs to industrial products to chemical formulae – the question of who will

    own what rights in the “output” of such programs may well become a hotly contested issue.’55

    Annemarie Bridy reflected upon the debate over copyright and artificial intelligence.56

    She charted the history of the discussion:

    For more than a quarter century, interest among copyright scholars in the question of AI authorship has

    waxed and waned as the popular conversation about AI has oscillated between exaggerated predictions

    for its future and premature pronouncements of its death. For policymakers, the issue has sat on the

    horizon, always within view but never actually pressing. Indeed, to the extent that the copyright system

    is now in a digitally induced crisis, the causes lie primarily outside the domain of cultural production,

    in the domains of reproduction and distribution. To recognize this fact, however, is not to say that we

    can or should ignore the challenge that AI authorship presents to copyright law’s underlying

    assumptions about creativity. On the contrary, the relatively slow development of AI offers a reprieve

    from the reactive model of policymaking that has driven copyright law in the digital age.57

    Bridy suggested: ‘AI authorship is readily assimilable to the current copyright framework

    through the work made for hire doctrine, which is a mechanism for vesting copyright directly

    in a legal person who is acknowledged not to be the author-in-fact of the work in question.’58

    James Grimmelmann has also considered the relationship between copyright, artificial

    intelligence, and robotics.59 He comments:

    Almost by accident, copyright law has concluded that it is for humans only: reading performed by

    computers doesn't count as infringement. Conceptually, this makes sense: Copyright's ideal of romantic

    readership involves humans writing for other humans. But in an age when more and more manipulation

    of copyrighted works is carried out by automated processes, this split between human reading

    (infringement) and robotic reading (exempt) has odd consequences: it pulls us toward a copyright

    system in which humans occupy a surprisingly peripheral place.60

                                                                55 Ibid. 56 Annemarie Bridy, ‘Coding Creativity: Copyright and the Artificially Intelligent Author’ (2012) 5

    Stanford Technology Law Review 1-28. 57 Ibid. 58 Ibid. 59 James Grimmelmann, ‘Copyright for Literate Robots’ (2016) 101 Iowa Law Review 657-681. 60 Ibid.

  • 20  

    Grimmelmann considers how the principles of copyright infringement and the doctrine of fair

    use deal with robotic readers.

    In 2017, Andres Guadamuz from Sussex University commented upon some of the

    implications of artificial intelligence for copyright law and policy:

    Creating works using artificial intelligence could have very important implications for copyright law.

    Traditionally, the ownership of copyright in computer-generated works was not in question because the

    program was merely a tool that supported the creative process, very much like a pen and paper. Creative

    works qualify for copyright protection if they are original, with most definitions of originality requiring

    a human author. Most jurisdictions, including Spain and Germany, state that only works created by a

    human can be protected by copyright. But with the latest types of artificial intelligence, the computer

    program is no longer a tool; it actually makes many of the decisions involved in the creative process

    without human intervention.61

    He suggests that the discipline raises fundamental doctrinal issues in terms of authorship,

    ownership, copyright subject matter, copyright infringement, and remedies.

    By contrast, the European Parliament Legal Affairs Committee has demanded the

    elaboration of criteria for ‘own intellectual creation’ for copyrightable works produced by

    computers or robots is demanded.62

    In the field of literature, there has been anxiety amongst authors and publishing

    regarding machine-learning projects of big IT companies. In 2016, authors expressed disquiet

    over Google using novels to improve its AI’s conversation ability.63 Erin McCarthy objected:

    It’s hard to gauge the use of my work and the exact purpose for its use without having seen it in action.

    My assumption would be they purchased a copy of the book originally. If they haven’t, then I would

                                                                61 Andres Guadamuz, ‘Artificial Intelligence and Copyright’, WIPO Magazine, October 2017.

    https://www.wipo.int/wipo_magazine/en/2017/05/article_0003.html 62 European Parliament Legal Affairs Committee, Civil Law Rules on Robotics that includes

    recommendations to the Commission on Civil Law Rules on Robotics (2015/2103/(INL); rapporteur, Mady

    Delvaux (S&D, Luxembourg) http://www.europarl.europa.eu/doceo/document/A-8-2017-

    0005_EN.html?redirect 63 Richard Lea, ‘Google swallows 11,000 novels to improve AI's conversation’, The Guardian, 28

    September 2016, https://www.theguardian.com/books/2016/sep/28/google-swallows-11000-novels-to-improve-

    ais-conversation

  • 21  

    imagine the source of the content, as intellectual property, should be properly attributed and

    compensated for the general health of the creative community.64

    Mary Rasenberger of the Author’s Guild maintained: ‘The research in question uses these

    novels for the exact purpose intended by their authors – to be read.’65 ‘It shouldn’t matter

    whether it’s a machine or a human doing the copying and reading, especially when behind the

    machine stands a multi-billion dollar corporation which has time and again bent over

    backwards devising ways to monetise creative content without compensating the creators of

    that content.’66

    For its part, Google was unapologetic about its use of the textual material:

    We could have used many different sets of data for this kind of training, and we have used many

    different ones for different research projects. But in this case, it was particularly useful to have language

    that frequently repeated the same ideas, so the model could learn many ways to say the same thing –

    the language, phrasing and grammar in fiction books tends to be much more varied and rich than in

    most nonfiction books.67

    No doubt the company was heartened by its victory over authors in a dispute over copyright

    and fair use in respect of Google Books.68

    In October 2018, a piece of AI art by the French collective Obvious was sold at

    Christie’s for $432,500. Hugo Caselles-Dupré of the collective Obvious explained that their

    work explored the interface between art and artificial intelligence. Their method relied upon a

    ‘generative adversarial network’. Caselles-Dupré explained:

    The algorithm is composed of two parts. On one side is the Generator, on the other the Discriminator.

    We fed the system with a data set of 15,000 portraits painted between the 14th century to the 20th. The

    Generator makes a new image based on the set, then the Discriminator tries to spot the difference

    between a human-made image and one created by the Generator. The aim is to fool the Discriminator

    into thinking that the new images are real-life portraits. Then we have a result.69

                                                                64 Ibid. 65 Ibid. 66 Ibid. 67 Ibid. 68 See - 69 Christie’s, ‘Is Artificial Intelligence Set To Become Art’s Next Medium’, Christie’s, 12 December 2018,

    https://www.christies.com/features/A-collaboration-between-two-artists-one-human-one-a-machine-9332-1.aspx

  • 22  

    He observed that ‘we found that portraits provided the best way to illustrate our point, which

    is that algorithms are able to emulate creativity.’70

    On March 2019, Sotheby’s is auctioning a work of AI art by the German artist, Mario

    Klingemann. Mario Klingemann has an unorthodox vision of creativity: ‘Humans are not

    original - We only reinvent, make connections between things we have seen.’71

    Certain AI Projects might also raise larger questions about copyright law, copyright

    exceptions, and copyright infringement. Montreal-based artist Adam Basanta has developed an

    AI project called All We’d Ever Need is One Another.72 Basanta has explained his process of

    using computer scanners:

    I was really surprised that the images looked a lot like canonical 1950s abstract paintings I literally had

    a moment where I had made a piece and I thought: I’ve seen this before. I looked it up and I found a

    Rothko that was very, very similar to it. “If it’s similar enough to a work that the art market or

    international collections has deemed art-worthy, then that image, which is similar to it, is also art-

    worthy. It becomes art.73

    The artist is being sued by Amel Chamandy for copyright infringement and trademark

    infringement of her paining, ‘Your World Without Painting’. 74 Her lawyer Pascal Lauzon said:

    ‘These acts of infringement illegally divert internet traffic away from NuEdge’s website and

    allows you to unduly benefit from the goodwill and reputation associated with the name and

    trademark AMEL CHAMANDY.’75 Amel Chamandy is seeking $CA 40,000 in damages.

                                                                70 Ibid. 71 Arthur Miller, ‘Can Machines Be More Creative Than Humans?’, The Guardian, 4 March 2019,

    https://www.theguardian.com/technology/2019/mar/04/can-machines-be-more-creative-than-

    humans?CMP=share_btn_tw 72 CBC Radio, ‘Can An Artist Sue An AI Over Copyright Infringement?’, CBC Radio, 13 October 2018,

    https://www.cbc.ca/radio/spark/409-1.4860495/can-an-artist-sue-an-ai-over-copyright-infringement-1.4860762 73 Chris Hannay, ‘Artist Faces Lawsuit Over Computer System That Creates Randomly Generated Images’,

    The Globe and Mail, 4 October 2018, https://www.theglobeandmail.com/arts/art-and-architecture/article-artist-

    faces-lawsuit-over-computer-system-that-creates-randomly/ 74 Ibid. 75 Ibid.

  • 23  

    Artificial intelligence has also been transforming music. Google’s WaveNet uses neural

    nets to generate speech and music.76 Stuart Dredge has reported that there are a number of AI

    firms, which seek forge links between the technology industry and the music industry.77

    Examples include AI Music, Amper Music, Popgun, Jukedeck, Humtap, Groov.AI, Magenta,

    and Flow Machines. Jon Eades – who runs the Abbey Road Red incubator - says that AI will

    have a mixed impact upon the music industry:

    I think there will be collateral damage, just like the internet. It created huge opportunity, and completely

    adjusted the landscape. But depending on where you sat in the pre-internet ecosystem, you either called

    it an opportunity or a threat. It was the same change, but depending on how much you had to gain or

    lose, your commentary was different. I think the same thing is occurring here. AI is going to be as much

    of a fundamental factor in how the businesses around music are going to evolve as the internet was.78

    AI music raises issues around copyright authorship, ownership, infringement, and exceptions.

    The law firm Allens has argued that there is a need to reform copyright law and policy

    to take into account artificial intelligence.79 Partner Andrew Wiseman worries: ‘The speed of

    progress and technology often overtakes legislation and that's been the case with copyright

    right back to when it first began with the Stationers in the United Kingdom and the 1600s.’ He

    contends: ‘I think a lot of people are not aware of how fast AI is working and proceeding, and

    the issue will be that if we don't deal with ownership issues like this quickly, then we may find

    that there is some very important work that ends up with no copyright because there's no

    relevant owner.’ However, others have protested that recognition of non-human authorship

    would go against the rationales and objectives of copyright law.

    There have also been parallel discussions taking place in respect of copyright law and

    non-human authorship in the context of the animal kingdom. The ‘Monkey Selfie’ case has

                                                                76 Devin Coldewey, ‘Google’s WaveNet uses Neural Nets to Generate Eerily Convincing Speech and

    Music’, Tech Crunch, 9 September 2016, https://techcrunch.com/2016/09/09/googles-wavenet-uses-neural-nets-

    to-generate-eerily-convincing-speech-and-music/ 77 Stuart Dredge, ‘AI and Music: Will We Be Slaves to the Algorithm?’, The Guardian, 6 August 2017,

    https://www.theguardian.com/technology/2017/aug/06/artificial-intelligence-and-will-we-be-slaves-to-the-

    algorithm 78 Ibid. 79 Lucas Baird, ‘Copyright Law Must Be Amended to Account for Artificial Intelligence’, Australian

    Financial Review, 1 January 2019, https://www.afr.com/business/legal/copyright-law-must-be-amended-to-

    account-for-artificial-intelligence-allens-20181227-h19hmb

  • 24  

    been quite notorious.80 The United States Court of Appeals for the Ninth Circuit held in this

    matter: ‘We conclude that this monkey—and all animals, since they are not human—lacks

    statutory standing under the Copyright Act.’81

                                                                80 Naruto v. Salter (2018) No. 16015469 81 Naruto v. Salter (2018) No. 16015469

  • 25  

    2. Commercial Law

    A. Problems associated with AI

    In a semi-fictional introduction to her book, Smart Technologies and the End(s) of Law: Novel

    Entanglements of Law and Technology,82 Mireille Hildebrandt, the Dutch lawyer and

    philosopher, outlined some of the applications of new technologies, including AI, in our not-

    too-distant future. Hildebrandt took currently available technology and applied it liberally to

    saturate Diana’s life in ubiquitous technological gadgets and applications to illustrate the

    control that these technologies, and the companies that own them, might have over our lives if

    we continue to thoughtlessly adopt technology. The ramifications for our human rights, our

    personal, social, and work lives, and our very existence are writ large in this story.

    Over the past ten or so years, applications of what is described as narrow Al have

    expanded and developed. Current applications that use powerful face recognition, data

    analytics and natural language processing have pushed AI further into our everyday lives. This

    has occurred at a speed and to an extent that we, as consumers, are unable to fully process. The

    implications for us continuing down this path, uncontrolled, are profound. Ordinary consumers

    of these products appear powerless against the technology companies. In its interim report on

    digital platforms, the ACCC found that:

    … consumers are unable to make informed choices over the amount of data collected by the digital

    platforms, and how this data is used. This reflects the bargaining power held by the digital platforms vis-

    à-vis consumers, and the information asymmetries that exist between digital platforms and consumers.

    The ACCC considers that the current regulatory framework, including privacy laws, does not effectively

    deter certain data practices that exploit the information asymmetries and the bargaining power

    imbalances that exist between digital platforms and consumers.83

    A number of the challenges to regulating developments in artificial intelligence (AI)

    and some possible solutions were set out by Michael Guihot, Anne F Matthew and Nicolas P

                                                                82 Mireille Hildebrandt, ‘Introduction: Diana’s Onlife World’ in Smart Technologies and the End(s) of Law:

    Novel Entanglements of Law and Technology (Edward Elgar Publishing, 2015). 83 Australian Competition and Consumer Commission, ‘Digital Platforms Inquiry: Preliminary Report’

    (Australian Competition and Consumer Commission, December 2018) 13

    https://www.accc.gov.au/system/files/ACCC%20Digital%20Platforms%20Inquiry%20-

    %20Preliminary%20Report.pdf.

  • 26  

    Suzor, in their paper, ‘Nudging Robots: Innovative Solutions to Regulate Artificial

    Intelligence’.84 This submission emphasises and develops several of those ideas. However, it

    also engages in and recommends a more commercial approach to regulation in this area. This

    is because, at a practical, commercial level, many of the possible harms associated with

    developments in AI stem from corporations pursuing profits through interactions with

    consumers. This commercial approach to regulating AI also fundamentally affects the choice

    of a suggested regulatory body to oversee the development of AI in Australia.

    Risk is a factor in determining regulatory responses. As discussed in ‘Nudging Robots’,

    the level of risk associated with various applications of AI is not constant. One application of

    AI (such as a facial recognition system) could pose a range of risks from low to moderate to

    high, depending on how it is used, by whom, and for what purpose. Further, the type of risk

    posed by each application may not be the same. For example, with a particular application of

    AI, there might be a low risk to safety or to human life, but a high risk of a breach of privacy.

    An additional complicating factor is that similar types of applications will be used differently

    in different industries or areas. For example, the same narrow AI application used in a product

    in medical procedures may be applied to a different product for security purposes. This will

    very likely mean that different regulatory agencies will be required to regulate the same AI,

    but in different applications. Therefore, it is too simplistic to seek to regulate it based upon a

    single presumed level of risk.85

    As argued in ‘Nudging Robots’, ‘public regulators must become informed about the AI

    used in their field, assess the risks posed by the AI application as it is used in the industry in

    which they operate, and regulate appropriately. Armed with a deeper understanding of the

    industry and the intended use of the AI, stakeholders involved in informing the regulatory

    approach will be better placed to ask the right questions to assuage, or at least contextualise,

    their concerns about levels of risk.’86 It is important in this context that there is a level of

    cooperation between regulating agencies.87

                                                                84 Michael Guihot, Anne F Matthew and Nicolas P Suzor, ‘Nudging Robots: Innovative Solutions to

    Regulate Artificial Intelligence’ (2017) 20(2) Vanderbilt Journal of Entertainment and Technology Law 385. 85 Ibid. 86   Ibid. 87 See Phil MacNaghten and Jason Chilvers, ‘Governing Risky Technologies’ in Matthew Kearnes,

    Francisco Klauser and Stuart Lane (eds), Crictical Risk Research: Politics and Ethics (John Wiley & Sons Ltd,

    2012) 99.

  • 27  

    B. Approaches to regulation

    Guihot et al canvass a number of approaches to regulating various incarnations of AI.88 The

    difficulty with a one approach fits all response becomes obvious when the nature of the problem

    is described. Matthew Scherer sets out four general ex ante problems with regulating research

    and development of AI: (1) discreetness, meaning ‘AI projects could be developed without the

    largescale integrated institutional frameworks’; (2) diffuseness, meaning AI projects could be

    carried out by diffuse actors in many locations around the world; (3) discreteness, meaning

    projects will make use of discrete components and technologies ‘the full potential of which

    will not be apparent until the components come together’; and (4) opacity, meaning the

    ‘technologies underlying AI will tend to be opaque to most potential regulators.’89

    These four problems go to the heart of the issues with regulating AI, but many other

    problems tied to individual uses of AI become apparent only after the AI is introduced to the

    public and its range of possible applications becomes apparent. For example, face recognition

    software is now readily available at low cost.90 The potential for abuse of this software should

    have been apparent, but its use in schools, social settings and, particularly, by government for

    ‘security purposes’ compounds the risks it poses and makes it an urgent problem in need of

    redress. Some possible regulatory solutions to developing AI are discussed below.

    C. Government intervention

    There is a role for government to play in regulating developments in AI. Nothing beats the

    effect of top down hard laws with enforceable sanctions to regulate behaviours. An immediate

    effect of legislation is the availability of remedies set out for its breach. A secondary impact of

    legislation is the deterrent effect it has on potential wrongdoers. In Ayers and Braithwaite’s

    articulation of responsive regulation, top down legislation is used as the ultimate enforcement

    in a pyramid of stepped regulatory interventions but also a deterrent when threatening heavier

                                                                88 Michael Guihot, Anne F Matthew and Nicolas P Suzor, ‘Nudging Robots: Innovative Solutions to

    Regulate Artificial Intelligence’ (2017) 20(2) Vanderbilt Journal of Entertainment and Technology Law 385. 89 Matthew Scherer, ‘Regulating Artificial Intelligence Systems: Risks, Challenges, Competencies, and

    Strategies’ (2016) 29 (2) Harvard Journal of Law & Technology 354-400. 90 Alex Walling, What Is the Best Face Recognition Software? (28 November 2018) Quora

    https://www.quora.com/What-is-the-best-face-recognition-software.

  • 28  

    sanctions. In this way “escalating forms of government intervention … reinforce and help

    constitute less intrusive and delegated forms of market regulation,”91 such as self-regulation.

    Some states in the United States have taken the lead against the use of biometrics such

    as face recognition. In a remarkably prescient example, in 2008, the state of Illinois legislated

    to prohibit its use.92 Under the Biometric Information Privacy Act, ‘no private entity may

    collect, capture, purchase, receive through trade, or otherwise obtain a person's or a customer's

    biometric identifier or biometric information’. The only exceptions are if the entity informs the

    subject in writing that biometric information is being collected, the purpose for which it is

    being stored, collected or used, and receives written release from the subject.93 A recent court

    case confirmed that the subject does not have to prove actual harm for the legislation to apply.94

    D. Self-regulation

    To be effective, any regulation must reflect the society’s norms. However, these norms are

    shifting as we become inured to the encroachments upon our privacy and intimate parts of our

    lives in return for access to social utilities such as provided by companies such as Facebook

    and Google. These companies, along with a number of others have set up a self-regulating

    entity called the Partnership on AI.95 The Partnership on AI’s purpose statement is to ‘benefit

    people and society,’96 and it is said to have been “[e]stablished to study and formulate best

    practices on AI technologies, to advance the public’s understanding of AI, and to serve as an

    open platform for discussion and engagement about AI and its influences on people and

    society.”97 It has developed a series of tenets, one of which is to seek to ‘maximize the benefits

    and address the potential challenges of AI technologies, by, among other things, … Opposing

                                                                91 Ian Ayres and John Braithwaite, Responsive Regulation: Transcending the Deregulation Debate (Oxford

    University Press, Incorporated, 1992) 4 http://ebookcentral.proquest.com/lib/qut/detail.action?docID=272606. 92 Biometric Information Privacy Act (740 ILCS 14/). 93 Ibid, s 15(b). 94 Russell Brandom, Crucial Biometric Privacy Law Survives Illinois Court Fight (26 January 2019) The

    Verge https://www.theverge.com/2019/1/26/18197567/six-flags-illinois-biometric-information-privacy-act-

    facial-recognition. 95 Partnership on AI Partnership on Artificial Intelligence to Benefit People and Society

    https://www.partnershiponai.org/. 96 Ibid. 97 Ibid.

  • 29  

    development and use of AI technologies that would violate international conventions or human

    rights, and promoting safeguards and technologies that do no harm’.98 In doing so, the

    Partnership on AI is taking on the role of a self-regulatory association and potentially warding

    off more enforceable state-imposed regulatory obligations. The benefits of this type of self-

    regulation are questionable. However, given the human rights concerns expressed by the HRC

    in relation to the behaviours and practices of these very institutions, it is not certain that this

    type of self-regulation will be wholly effective.

    The commercial applications of AI are becoming more innovative and pernicious. The

    world's most valuable companies99 are investing heavily in AI’s potential, partly because of the

    immense sums of money to be made. For this reason, it would be problematic to allow these

    companies to offer the solutions to the problems with AI. These technology companies

    pursuing corporate strategies at the expense of consumers has resulted in what Shoshana Zuboff

    describes as Surveillance Capitalism. Zuboff defines Surveillance Capitalism variously as:

    1. A new economic order that claims human experience as free raw material for hidden commercial

    practices of extraction, prediction, and sales; 2. A parasitic economic logic in which the production of

    goods and services is subordinated to a new global architecture of behavioural modification; … 8. An

    expropriation of critical human rights that is best understood as a coup from above; an overthrow

    of the people’s sovereignty.100

    For these reasons alone, self or co regulation by these entities may be one strand in the

    regulatory rope, but cannot stand alone.

    E. Nudging

    As discussed in ‘Nudging Robots’, nudge theory has become prominent in recent years.101

    Guihot et al argue that:

                                                                98 Ibid. 99 TOP 10 - The Most Valuable Companies in the World - 2019 List | FXSSI - Forex Sentiment Board

    https://fxssi.com/top-10-most-valuable-companies-in-the-world. 100 Shashona Zuboff, The Age of Surveillance Capitalism: The Fight for a Human Future at the New

    Frontier of Power (Hatchet Book Group, 2019) i. 101 See, e.g., id. at 6–8.

  • 30  

    Psychological observations, as applied in behavioral economics, reveal that normative human behavior

    can be skewed or distorted by inherent human biases. Nudge theory proposes that by exploiting these

    biases, humans can be nudged to behave so as to achieve an outcome desired by the nudger. The theory

    has tended to focus on nudging individual behaviours. However, recent work has examined how

    behavioral economics approaches might influence a broader spectrum of decision-makers—including

    corporations and policy-makers.102

    It should not go without saying that the role that the Human Rights Commission in its

    Human Rights and Technology Project (HRTP) and now the HRC with the World Economic

    Forum (WEF) is undertaking is itself nudging behaviours of developers in AI. Discussing and

    documenting the human rights aspects of technology development brings these issues to either

    the conscious or the sub-conscious of those exposed to the HRC and WEF work. The effect of

    this is unknowable, but should not be dismissed simply because of that.

    F. Regulation by Design

    Lawrence Lessig envisioned regulation as a combination of the forces applied by law,

    architecture, markets and social norms. He popularised the idea that the architecture of

    computer code could be a regulatory tool.103 That is, that computers could be coded so as to

    self-enforce a legal outcome, or to prevent a non-legal outcome. In this way, computers and

    the internet could be self-regulating. Lessig noted that:

    We can build, or architect, or code cyberspace to protect values we believe are fundamental, or we can

    build, or architect, or code cyberspace to allow those values to disappear. There is no middle ground.104

    This notion has led various academics to theorise on a number of issues that could be governed

    using designs or architecture of computer code. These areas now include Privacy by Design

    (PbD), Data Protection or Security by Design (SbD),105 and Legal Protection by Design (LPbD).

                                                                102   Michael Guihot, Anne F Matthew and Nicolas P Suzor, ‘Nudging Robots: Innovative Solutions to

    Regulate Artificial Intelligence’ (2017) 20(2) Vanderbilt Journal of Entertainment and Technology Law 385. 103 Lawrence Lessig, Code and Other Laws of Cyberspace (Basic Books, 1999). 104 Ibid 6. 105 See Lee A Bygrave, ‘Data Protection by Design and by Default: Deciphering the EU’s Legislative

    Requirements’ (2017) 4 Oslo Law Review 105 (discussing Article 25 of the GDPR on Data Protection by Design).

  • 31  

    In his book, Privacy’s Blueprint, Woodrow Hartzog emphasised design as a way to

    defend against invasions of privacy. He argued that

    Design-based protections can require companies to protect against reasonably anticipated third-party

    harms. Technology users will not have to worry as much about hiring an attorney or suffering only to

    receive a paltry recovery, because they will not become victims. Design cannot do everything, but it

    can dissuade would-be wrongdoers if the transaction cost is high enough.106

    Hildebrandt proposed Legal Protection by Design (LPbD), in the same vein as the arguments

    put forward for Privacy by Design and Security by Design. Hildebrandt argued that:

    LPbD, then, requires that the legal conditions we have agreed upon (articulated in natural language) are

    translated into the technical requirements that inform the data-driven architecture of our everyday

    environment. These requirements should instigate technical specifications and default settings that —

    other than current systems — afford the protection of fundamental rights. Thus, LPbD should constrain

    the data-driven architectures that run our new onlife world, while challenging developers to offer multiple

    ways of modelling the world, thus making their calculations, predictions and pre-emptions both testable

    and contestable. Instead of ‘anything goes’ for the architects of this new world, democratically

    legitimated law must regain its monopoly on setting the defaults of societal order, defining the rules of

    the game in a way that brings the data-driven machinery under the Rule of Law.

    When it comes to regulating AI, there is no one ‘AI’ and there is therefore no corresponding

    single regulatory response. There must be a multi-dimensional response using elements of

    government rule, self-regulation, design aspects and other soft regulatory tools (we use nudging

    as an example of this). Certainly, an emphasis on the design approach to AI development, could

    be an effective tool in the regulatory arsenal to govern developments in AI.

                                                                106 Woodrow Hartzog, Privacy’s Blueprint: The Battle to Control the Design of New Technologies (Harvard

    University Press, 2018) 82 http://ebookcentral.proquest.com/lib/qut/detail.action?docID=5317538. However, see

    Bert-Jaap Koops and Ronald Leenes, ‘Privacy Regulation Cannot Be Hardcoded. A Critical Comment on the

    “Privacy by Design” Provision in Data-Protection Law’ (2014) 28(2) International Review of Law, Computers &

    Technology 159 ('The upshot of our analysis is that ‘taking privacy seriously’ is unlikely to be achieved by

    focusing on rule compliance through techno-regulation, which leads system developers into a morass of selecting

    and translating rules that cannot be simply translated into system design requirements. Instead, taking privacy

    seriously requires a mindset of both system developers and their clients to take into account privacy-preserving

    strategies when they discuss and determine system requirements.’).

  • 32  

    G. A Commercial/Consumer Approach

    Several of the submissions to the HRC report lamented that Australia lacks a bill of rights that

    would ensure human rights were incorporated into Australia’s system of governance. In the

    current political climate, it is difficult to envisage such a development. However, there is

    another broad-reaching, almost universal, regime that is already in place that does act in the

    best interests of the vast majority of Australians. The Competition and Consumer Act 2010 and

    the Australian Consumer Law (ACL) apply to interactions with consumers. Under the ACL, a

    person is taken to have acquired a good or service if the cost of the good or service did not

    exceed $40,000 or the good or service was a kind ordinarily acquired for personal, domestic or

    household use or consumption.107 This broad classification of transactions as consumer

    transactions would also arguably capture many of the goods and AI services provided by

    technology companies. If it does not, then this submission argues for amendments to the ACL

    to include such interactions.

    Indeed the breadth of the application and protections sought to be covered in the ACL

    was envisaged at its inception in 1974 when the then Trade Practices Bill was being debated.

    The Second Reading speech of William Morrison, the then Minister for Science, on the

    introduction of the Trade Practices Bill in 1974 [at pp 574-575] is worth quoting at length. The

    Minister said:

    I think we all know that consumers are the largest but regrettably the least organised economic group in

    the community. Every one of us, by definition, is a consumer—from when we get up in the morning and

    squeeze our toothpaste tube until we go to bed at night and turn off the lamp. But more often than not,

    we are quite ignorant of our rights and privileges as consumers. What we are proposing in this Bill is

    a consumers' charter, that is, a bill of rights for Australians as consumers. We, as consumers, have

    first the right to be safe, the right to protection against products which could harm our health or endanger

    our lives.

    Secondly, we have the right to know. The march of technology has brought added difficulties

    as well as benefits to the consumer. The housewife today is required to be an amateur mechanic,

    electrician, doctor, chemist, food technologist and mathematician, but all too rarely is she given the

    information she needs to fill these roles. Our consumers' charter gives her the right to have access to the

    facts, free from fraudulent or misleading information, whether in packaging or advertising.

    Thirdly, consumers should have the right to choose, the right to select between goods and

    services at competitive prices. Our charter protects the consumer against shady business practices which

                                                                107 Australian Consumer Law s 4B.

  • 33  

    restrict the basic right to get value for money. ... To recapitulate, related to this Bill and involved in this

    Bill are four fundamental rights: The right to 'be safe, as proposed in clause 62; the right to know, as

    proposed by clause 63; the right to choose, and the right to be heard.

    And at [575-576], the Minister said:

    I turn to the approach adopted in the Bill. The Bill prohibits a wide range of specific practices which

    have come under notice, including false representations, bait advertising, referral selling, pyramid selling

    and the assertion of a right to payment for unsolicited goods. It is not possible however to specify in

    advance the nature of all undesirable practices, as sharp operators continually evolve new schemes

    for duping the public. For this reason the broad prohibition of misleading or deceptive conduct in

    clause 52 is of great importance. … The courts will be able under that provision to take action against

    conduct which may not fall within the more specific terms of other provisions. This will provide the

    flexibility necessary if legislation of this kind is to be able to deal with evolving market practices without

    the constant need for legislative action to catch up, often after much damage has already been done, with

    new practices that are harmful to consumers.

    The audacity and prescience of this legislative proposal is impressive. It anticipated

    technological advances, and sought to address them. However, it could not have predicted the

    extent of technological advances in the last twenty years or so. To frame the legislative scheme

    as a bill of consumer rights speaks to how broadly it applies and its potential impact in the area

    of AI development.

     

  • 34  

    3. Regulation

    There have been a number of national inquiries and discussion papers in respect of robotics

    and artificial intelligence.

    A. Australia

    The Australian Parliament held an inquiry into drones in 2016, and into autonomous vehicles

    in 2017. The Australian Centre for Robotic Vision produced A Robotics Roadmap for Australia

    in June 2018. Synergies Economic Consulting produced a consultancy report on The Robotics

    and Automation Advantage for Queensland. The Australian Human Rights Commission has an

    ongoing inquiry into human rights and new technologies – considering amongst other things,

    robotics, artificial intelligence, and the Internet of Things. The law firm Corrs has stressed: ‘It

    is time for Australian law makers to start thinking about how we want our life with robots to

    be regulated’.108

    In this context, the Australian Human Rights Commission has proposed the

    establishment of a Responsible Innovation Organisation.

    The scope of this organisation is unclear in the white paper. There is a slippage between

    the discussion of artificial intelligence and the new institution in the white paper. It is unclear

    whether the Responsible I