logo
Call us: +44 (0)1865 244727

  • Home
  • Scope
  • News
  • Products
    • RADAR
    • CALL-OFF PROJECTS
  • Clients
  • Contact
  • How we work
    • Independent
      • Common law orthodoxies
      • Sensationalism
      • Expert witness
      • Regulation and Politics
      • Tied services
    • Up-to-date
      • Timely
      • Insurance Scenarios
      • Probabilistic Methods
    • Expert
      • Personal Injury
      • Trends
    • Innovative
  • Database
    • Member’s login
    • Member’s Settings
    • Register
    • RADAR Database
  • Recent projects
    • EMFs
    • STRESS AT WORK
    • WHIPLASH
    • WELDING RODS: MANGANESE EXPOSURE
    • ENVIRONMENTAL TOBACCO SMOKE
    • Other Projects



Artificial Intelligence – who to blame?

Feb 13, 2014
by Andrew@Reliabilityoxford.co.uk
0 Comment
Following a lecture at the Oxford Martin School; 13th Feb 2014.

Task specific machines have been with us for a long time. Sometimes they go wrong. Who to blame? Well it is worth pointing out that slaves and employees have been around for a long time, so liability rules have become quite predictable. A machine is just the agent of the master/owner/user no matter how “intelligent”.

What’s different about AI?

In AI the machine develops and refines its own optimum manner of performing a function. Of course it does this within the limits of its methods of action, the feedback it gets from taking those actions and the efficiency of the optimisation software.

Provided the system is well designed, then the scope of potential harms can be factored out. The designer is to blame. If the system is misused, the user is to blame.

But

Utility maximisation is not always the social optimum. A self-driving car might find the optimum route but it is a route which would be ethically unacceptable. Where ethical choices are to be made, the designer must ensure that ethical optimisation is included. Again, the designer is to blame. Being law-abiding may not be sufficient – there is no comprehensive law.

But

Given enough power, some AI machines might in effect, design themselves and therefore have quite unpredictable emergent misbehaviours and even an ethical system of their own. Society should perhaps have some say in whether self-design is allowable and if so how much latitude should there be.

 

Social Share

Leave a Reply Cancel reply

*
*

captcha *

Search Documents


Categories

  • Causation
    • de minimis
    • material contribution
  • Date of knowledge
  • Diagnosis
  • Duty of Care
  • Exposure estimation data
  • Mitigation
  • Motor related injury
  • News
  • Uncategorized

Archives

  • July 2023
  • June 2023
  • November 2020
  • January 2020
  • November 2019
  • May 2019
  • April 2019
  • March 2019
  • January 2019
  • December 2018
  • November 2018
  • September 2018
  • August 2018
  • July 2018
  • June 2018
  • April 2018
  • November 2017
  • July 2017
  • April 2017
  • May 2016
  • April 2016
  • November 2015
  • April 2015
  • March 2015
  • December 2014
  • October 2014
  • July 2014
  • April 2014
  • February 2014
  • December 2013
  • November 2013
  • October 2013
  • September 2013
  • July 2013
  • June 2013
  • April 2013
  • March 2013
  • February 2013
  • January 2013
  • December 2012
  • November 2012
  • October 2012
  • September 2012
  • August 2012
  • July 2012
  • June 2012
  • May 2012

© Re: Liability (Oxford) Ltd. 2012. All rights reserved.
Website Design by The Big Picture