Being that this is a computer, it's in a department I'm
comfortable and familiar with. My experience with human rights
is mostly limited to a youth rights activist thing I did back in
1990-1997 on the early Internet - it spawned a few more lobbying
type groups, but my interest was more freedom of
expression/creativity/education rather than voting rights,
parenting rights, etc.
Anyway - being that it's a computer... there's a lot of people
involved. This tree illustrates the everyday issues with making
anything.
A flaw in any part of the process and you end up with a
half-assed product. This is the norm rather than the exception;
problems with robot bomb sniffers were notorously awful in the
beginning stages. Nobody was listening to anybody.
Anyway, assuming it makes it to a compromise where everybody is
happy enough with the product (in this case, a flying computer
that follows the instructions given to it ahead of time by the
users, the engineers, the programmers, etc.
So, is there a human in charge? Yes. There is *many* many humans
in charge, and they all play a part in making something like
this work according to plan.
Or at least try to. The drone they end up with will be
all-too-human, with all of the flaws programmed fully into it.
How they will use it? That's the territory you're talking about.