FREE TOGETHER

FREE TOGETHER

Don’t bring me your mythology
Your clapped out ideology
For I can watch your wars no more
Don’t quite know what a god is for
Except to drug the people’s mind
With one true way, an only kind,
Of being young or being old
Whilst others stuff our mouths with gold
To ensure we do what we’re told
For if we don’t they soon will scold,
Not quite enough to set us free,
You can’t be you, I can’t be me.
Dont need fake beings up above,
Together, all we need is love
In sixty seven John was right
There really is no need to fight.
So you be you and I’ll be me,
Together set each other free.

Harry Rogers, In the Yellow Room, 14th July 2020.

WE WANT FRIES WITH OUR MEAL DEAL.

Steer Kharma, forensick, at despatch box,
Mewls as Haystacks gang shoots fox after fox.
Cummings has stolen all Jeremy’s clothes,
To wave them beneath New New Labour’s nose.
Forests of money trees bloom at the bank,
Quantitavely eased with clink and clank.
Billions of pounds are drawn at a stroke,
Millions of workers now left for broke.
Advisory rules now go up in smoke,
Health ministers stats now called out a joke.
The track and trace app? A pig in a poke,
Changes in benefits soon to revoke,
Dole queues grow longer, this mess is severe,
Still, grab a meal deal, don’t say it’s austere.

Harry Rogers in the Red Bedroom, Friday 10th July 2020

MESSAGE TO MICHAEL IN TUCKER, GEORGIA, USA.

Here, in the disunited fiefdom, where a man with what looks like a storm blown stook of straw on his head rules the roost, us mere mortals have been offered a meal deal instead of a new deal. Up to ten pounds a punter to cover 50% of the cost of eating a meal out every Monday, Tuesday, and Wednesday as a way of enticing us all to get back into socialising whilst at the same time saving the businesses of restauranteurs and publicans. Meanwhile Covid spikes all over the place and we learn that indoors two metres is not much of a defence against an airborne virus. Jenny and I are staying right here harvesting our raspberries, weeding the vegetables and reupholstering the old sofa bed. Still, the muse of the iambic pentameter is ever present as the sonnets pour out of my fingers and into my phone at an alarming rate, and I am surprisingly jolly.
The birds are as busy as ever, swallows and swifts swoop over the hillside lunching on the wing and woodpeckers use our nut feeder as a crazy kind of swing. Life is precious. One love, companero.

PANTO ON FIRE

A nest of incestuous investors
Lies behind cloak of bombed out BJ
Hand grenade drops through Brexit pill box slit
Wounded spads analyse pin puller disguise
Desperate to find who will benefit
Us cannon fodder voters, smoke in eyes,
Watch in stunned terror as panto unfolds,
No-one shouts Look Behind You, in the wings
Waits latest parvenu, heart all a quiver
Soon, on centre stage, where he will slither,
Highwayman new shouts Stand and Deliver
Fresh spads snivel as the people shiver.
Democracy turns to patrician mauve
I really do hope it’s not Michael Gove.

Harry Rogers, in the red bedroom, Monday 25th May 2020.

CHOMSKY’S PEKINESE

Chomsky’s Dog chews papers in the background
Every now and then makes a growling sound
Naom proselytises without pause
Dog scratches purposefully with all claws
Advice for activists flows out freely
Words spoken softly yet no less steely
Offers hope for future generations
Twenty years to save the fate of nations
Wretched theives and crooks, wrecked economy
Post Covid climate, lockdown anomie,
Our world in danger, soon we will be toast,
All now take action, don’t give up the ghost.
He is compelling, get up off our knees
Shred Tory lies like Chomsky’s Pekinese.

Harry Rogers, in the yellow room, Pencnwcau, 24th May 2020

NO MORE WILD SIDE

NO MORE WILD SIDE
There is no wild side to walk anymore
All is normal now, we all know the score.
Sky God worshippers lay down holy law
Virus, hurricanes, capitalist war
Perfect storms rage together, globally
No wild side to walk, not for you and me
This is how it ends, locked in misery
No wild side to walk, not for you and me.

Lou said
Hey babe
Take a walk
On the wild side

Hey babe
Lou’s dead
Walk the talk
No more wild side

Wild Side No More

Wild Side No More

All normal now
All normal now

Wild side?
No More!

Harri Rogers
In the yellow room,
Pencnwcau, Aberbanc
28th February 2020

NO STONE ROLLING BACK

No stone rolling back across front of tomb,
The land of hope has run right out of room
No resurrection allowed this time round
No angels arise above hallowed ground
These modern Neo liberal Romans
Busily launch their shiny slick showmen,
Believe they have cracked it once and for all
Tear down our leader and smile as he falls
Crucify disciples, don’t hesitate,
Nullify their love, replace with fake hate.
Spinners turn his words inside upside down
Whilst drooling sheep vie obsessed with The Crown
In another place crooks get something done
Turn taps on fully, soon the pound will run.
J C, warns like so many times before
Of rabid Etonian dogs of war.
Nobody listens, the few rule alone,
Rise quickly, we need a new rolling stone.

Harri Rogers in my red bedroom, Aberbanc 16th February 2020.

Boomtime In Dystopia by Simply Spiffing

In November 2019 my friend from Atlanta Georgia, Chris Ricker, came to Wales for 22 hours. Whilst he was here with his buddy Sean from Florida we recorded this reworking of my lyric Boomtime In Dystopia at LTS studios in Llanon.

Listen to Boomtime -Simply Spiffing by Harry Rogers on #SoundCloud
https://soundcloud.com/harriboy/boomtime-simply-spiffing

FROM HEARTBREAK TO HOPE

A poem written as we began canvassing in the 2019 election campaign. Seems kinda prophetic now.

From heartbreak to hope as losers we row,
On past winners boat where slumped bodies loll,
Tomorrow another race day beckons,
Tactics change, different strokes to be pulled,
Cox chooses channel, allays all our fears,
Our blades cut the water at course’s end
Row on past winners, once more we must pull,
On down that river from heartbreak to hope.
Way down,
Way way down
From Heartbreak to Hope
Row on, on, on,
Row on,
From Heartbreak to Hope.
Never stop rowing
From Heartbreak to Hope.

Harri Rogers
Pencnwau, Aberbanc
19th November 2019

BURNT FEATHERS

A sonnet from home.

BURNT FEATHERS

I smell burnt feathers adrift on the air,
Mingled with sharp seasoned, iron blood everywhere.
Fires roar round L.A. , deluge tears across town,
Sidney burns,almost, Fukushima frowns.

Acrid taste endures, pain insane, rain blame.
Not my fault, or yours, Johnson mops, plays game!
Inhale burger weed flavour on street breeze,
Weekender – London, beggars on their knees.

Miserable band cuts straight through to me,
Lone trumpet soaring over red blue screen.
Anyone would think, with stench all around,
We might waken up, find some common ground.

But no, drink warm gin, ginger lemonade,
World goes up in smoke, burnt feathers pervade.

Harri Rogers
In the hut
11/11/2019

Tippy Toe To The Boogie – episode one Superman’s Belt Buckle

I published my new episode Superman’s Belt Buckle Mar 26, 2019 17:18, please check it out
https://www.podbean.com/media/share/pb-ge8nz-ac0b55

https://www.podbean.com/media/player/multi?playlist=http%3A%2F%2Fplaylist.podbean.com%2F4568366%2Fplaylist_multi.xml&vjs=1&kdsowie31j4k1jlf913=367408e87db25b468adfde17a5a48d3dc6d57ccb&size=240&skin=7&auto=0&download=1&pbad=1

The Aberbanc Argus 11/11/2018.

Armistice Day 2018, the centenary of the end of WW1. What a strange day. The usual dirge like tones on the BBC Radio Four coverage of the eleventh hour on the eleventh day of the eleventh month as the wreaths were laid at the cenotaph. Not long afterwards the MSM swung into action and launched the inevitable attack on Jeremy Corbyn for wearing the wrong kind of coat and too small a poppy. Surely everyone now sees this kind of tabloid attack for what it is, cheap, shoddy and laughable. As he stood, surrounded by a bevvy of haute coutured war mongering murderers responsible for untold numbers of atrocities, as a man who has spent his whole life supporting peace campaigns he must have revelled in the irony. Of course the size of ones poppy is far more important than sanctioning millions of deaths in Iraq, Afghanistan, Libya. Also, having ones rain hood out is a much bigger insult to the dead than those politicians who continue to broker massive arms contracts whenever and wherever they can across the globe. Such hypocrisy is startling but this is after all a symptom of the fear felt by the establishment of a Jeremy Corbyn led government. Meanwhile in Paris French leader Macron publicly slaps Trump in the face by lecturing on the difference between nationalism and patriotism whilst in Warsaw a massive demonstration orchestrated by Nazis is attended by the head of state and other government politicians. All this whilst I sit constipated and convalescing from a hernia operation last Tuesday that is ameliorated by codeine and paracetamol. I read in The Sunday Remainer (The Observer) that the MOD are now developing drones that can take autonomous decisions on whether to kill or not, something I have been writing about for the last thirteen years, ironic on the centenary of the end of the war to end all wars that new fangled killing machines are being developed for the very self same politicians who wear the larger poppies, believe in nuclear weapons as deterrence, and are draped in tailored mourning clothes. After a plate of frozen berries ( defrosted), muesli and plain yoghurt at lunchtime I finally have a shit after six days of discomfort, but I still feel sickened by the historical stench of wars past mixed with the fear that something awful is brewing in the very near future. All this and Charlton Athletic scored away from home in the FA Cup first round at Mansfield Town thus earning a draw and a place in the second round draw. See, I told you it was a strange day. Sleep well comrades, whilst you can.

Nowhere To Go Go

Caught short in Tokyo?
No problem a go go.
Public loos never shut,
Designer works of art,
Not like that Nissan hut,
Pugged away, kept in dark.
Most in convenience
Drab, ugly, plain and stark,
Squalid where we spend pence.
Unlike the Japanese
Who all luxuriate
With fancy poos and pees,
Their toilets truly great,
Aesthetic and pleasing
Built like finest palace,
For farting and easing,
Not a poisoned chalice
Like those cold windy sheds
Built of tin and concrete,
Crass stainless steel piss heads,
No more found on high street.
Suppose we’ll carry pos
When all the loos are gone.
Where cherry blossom grows
Lavatories live on!

Harri Rogers, Pencnwcau, 1st June 2018.

The Repository Of Socially Useful Ideas.

This is an idea that came to me in a dream whilst on a Christmas holiday in Palma on 27th December 2015.

I was dreaming that I was running a kind of Blog. A blog where people could lodge their ideas of how to build a better world. The blog would have a portal a bit like the entrance into Narnia. Once through the entrance a repositer would have access to all the ideas lodged there. There would be open access to ideas and they would be lodged by category.  Unlike Wikipedia, which is an attempt at a self monitoring encyclopedia, this would be a repository where just ideas are posted. At the moment there are many academic websites that cost an arm and a leg to accessand are beyond the reach of the vast majority of people who neither earn the salaries nor have the economic backing of major institutions. Also, the current social media are great places to chat and organise events but when it comes to the sharing of ideas they fall down big time.  This is mainly because they are not designed with a coherent archival capability.  The website Democracy Now comes close to doing this but has as it’s central aim the dissemination of left wing news and keeps a magnificent global archive of news events.

What I am proposingis a modern day version of The Left Book Club where strategic and political thinkers and polemicists can rub shoulders with activists who have great ideas.

The need for security would be an absolute imperitive. To this end I believe that something akin to a block chain would need to be used in order to keep the maliciously intended from subverting or riddling the site with virusus or software that could be used by governments to access lists of contributors. 

Contributors would adopt an avatar/identity at the start of their journey into the repository and would be able to access all ideas posted without hindrance.

The blog will need very careful monitoring and a secure vetting process to ensure that neither political nor religious factions can subvert or occupy the site for their own ends.  The overarching purpose of the site is to build a repository of knowledge that is open to all. A bit like what Tony Benn often refered to as a people’s university. There are still public libraries and reading rooms but these are increasingly disappearing as funding for the reproduction of traditional books, the cost of maintaining and heating the often large buildings, and cost of employing librarians and other staff, all become prohibitive.

What I am proposing here is an online space where people can lodge ideas in the form of extracts, essays or longer tracts. This would be a place where these ideas would be given for free. I realise that for many academics this is anathema. A seismic blow to the concept of selling intellectual property. However this has already happened to a large extent in the creative industries where streaming is trying to outgun filesharing.  I am suggesting that ideas are not the property of those that have them but rather they are the product of the world we inhabit and as such should be shared. Zut alors I hear you say, this is the end of the world for so many people. What about copywrite? How can academics make a living? Well the key here is that people would donate their ideas to the repository freely and of their own volition.  There would be no expectation of payment and no exploitation through sales. The ideas would be there to be used, studied, or enacted.

A site like this would need some initial financeand also some ongoing revenue to cover any on-costs. This could come from donations, gifts and fundraising events.

The main idea though is to get the realm of academic freedom away from the chaos that is Facebook etc and into a modern online space where ideas are freely exchanged, tested, challenged and developed. TED does this up to a point but what is needed is an online place where anybody can deposit ideas that they believe are for the common good. Where ideas can be disseminated and shared with the sole purpose of making the world a better place. At the moment our Universities are, in the main, set up to compete and generate finance through competitive business models but this is not the only or even the best way things can be organised.

We need another way. We need a place where anyone who has a good idea can share it, not for money but out of altruism. The creators of Facebook, Google, Microsoft, Apple, etc etc have taken the internet away from the people,monetized it to the nth degree and have handed back a cyber lawyers paradise where the original idea of the internet has become subverted and enslaved to mammon. Bitcoin, despite it’s faults offers a possible way out through the use of blockchain technology. I am not an expert on this but I believe it is possible to set up an online entity that exists free from the constraints and exploitation of capitalism true to the original concept of the world wide web. It is something we need to strive for and a good start could be the creation of The Repository of Socially Useful Ideas.

Merthyr Rising Festival 2018 – Red Poets Events.

​I am doing two readings with The Red Poets at the Merthyr Rising Festival.
We’re doing 2 sets on Saturday May 26th.
The first is at Theatr Soar, just off High St.  Max 2 poems each and we’re on from 11 – 12.
Second is in the tent in Castle Car park, near Soar and we start at 13.15. Again 2 poems each.
This will be the running order for both sessions –
Mike Jenkins 

Des Mannay

Gemma Howells

Tim Evans

Heather Pudner

Phil Howells

Al Jones

Julie Pritchard

Tim Richards

Mike Church

Harry Rogers

John Williams

Andrew Bartz

Rhoda Thomas 

Phil Knight

Heather falconer

Rhys Milsom

Huw Pudner

Rob Cullen

Patrick Jones ( tent only)

Barry Taylor.

Windflower – Poem 8th May 2018.

Windflower

The seed arrived
Without warning
On an unknown
Foreign Zephyr.
Depositing
Itself, neatly,
Between dry stones.
On spagnum green
Softly nestled
For duration
Of summer warm
Swollen with dew
Bursting upwards
Searching for sky
Seeking sunshine
Stalkly groping
Stronger each day
Budly bursting
Cerulean
Bluely special
Shiny dawning
Unexpected
Glory morning
My windflower

Harry Rogers: Tea shop in Newcastle Emlyn, 8th May 2018

Unbroken Ponies – Tippy Toe To The Boogie video.

Recorded in Andy’s Gaff studio in Frome last November with Steve Young – Guitar, Robert Goldsmith – Saxes, Andrew Howell – Bass and Drums, Harry Rogers – Vocals.  This little earworm is all about a festival I went to in USA last summer.  I wrote the words sitting in Dr Bombay’s Tearoom in Candler Park, Atlanta, an excellent place to sit and get in contact with your muse.

NEW NAME, SAME ME

I have adopted a one letter change to my name for all my published work, online and elsewhere and for my internet presence from Harry to Harri.  This is because I have lived longer in my current home in Wales than anywhere else throughout my 69 years.  My email handle is already harriboy so it is a small step but one that puts me more in tune with the country I live in.  I will still be writing in the same vein as ever and hopefully will be updating this blog more often.

The other reasons why Labour lost in 2015

Think Left

For the most part, Margaret Beckett has managed to avoid the firing line for her 35 page report as to why Labour lost the 2015 GE.  Essentially, the report (which can be read here) does not fit easily into the Labour Right’s or the media’s frame of reference… vague or bland was the best they could come up with.  The press tried to whip up some excitement about ‘the suppression of a secret report’ about focus group findings but the task of blaming Jeremy Corbyn for Labour’s defeat in 2015 eventually proved too convoluted.  However, Jamie Reed MP did his best in a valiant effort for Progress:

Any Labour leader who refuses to listen to the country and who prizes the views of Labour members above Labour voters and former Labour voters will likely find that although they may secure the Labour crown, they will lose…

View original post 1,583 more words

What do the Labour Right think will happen if ‘they get their party back’?

Think Left

To be honest, I’m a bit mystified by the pronouncements of the Labour ‘modernisers’. Their thinking doesn’t seem to make any sense.

In the last two weeks, since Jeremy Corbyn won the Labour leadership with 59.5% of the vote, we’ve heard Charles Clark say “It’s obviously been a very bad week for him.” (?) and that Yvette Cooper is waiting in the wings to take over as Labour leader if Jeremy Corbyn gets fed up with the job (?).

Much more ominously, there are ‘reports that elements of the British Army may rebel if Jeremy Corbyn is elected Prime Minister, and a serving general quoted as saying that“The Army just wouldn’t stand for it”.

Has it been a bad couple of weeks for JC?

Off the top of my head, he has put together a huge and an inclusive shadow cabinet, which comprises 52% women; made a very…

View original post 1,224 more words

Working Together Against Drone Terror.

Speech for Porthmadog – 15th November 2013

Cyfeillion, a’i gyfeillion, noswaith dda a diolch i chi am fy ngwahodd.

Friends, and comrades, good evening and thank you for inviting me. I apologise for not being able to address you in Welsh tonight, I am afraid you will have to put up with English for the rest of my speech.

My name is Harry Rogers I bring greetings from Bro Emlyn For Peace and Justice, a local activist group that formed in 2003 as a response to Tony Blair’s illegal war on Iraq and we have been involved in a number of campaigns since that time. I personally have taken a keen interest in the research, development, and deployment of unmanned aerial vehicles (UAVs) and in particular Unmanned Combat Aerial Vehicles (UCAVs) since about 2006.  I have spoken on a number of occasions to this subject and the last time I spoke at a public event on this issue was at an anti drone conference in Cardiff earlier this year.  That conference led to the formation of Drone Campaign Network Cymru which is  umbrella organisation co-ordinating information about, and activities against military & surveillance drones in Wales & beyond and I am now a member of the co-ordinating committee of that group.  We recently organised a very successful protest demonstration outside the entrance to West Wales Airport at Parc Aberporth which is where the MOD are carrying out the airworthiness testing of the Watchkeeper drone, or at least this is the official line, I have my doubts on this and will touch on some of what I suspect is happening there later on in this speech.

I want to address a number of issues tonight and these are:

– A brief history of recent events at Parc Aberporth and the MOD base at Parclyn

– What is the current state of play vis a vis drone technology?

– Why is there such acquiescence in the use of Drones by politicians from all the major political parties?

– How do we move forward to bring about an end to the development, testing and deployment of drones in Wales and beyond?

I will take questions at the end of the speech so please do feel free to ask me anything you want on what I have said and also ask my view on any ideas that occur as a result of listening to the talk.

A brief history of recent events at Parc Aberporth and the MOD base at Parclyn

I suspect that you already know quite a lot of the old history relating to RAF Aberporth and RAF Llanbedr in that they were both used to test missiles up to both bases being closed down.  Everybody in West Wales and in North Wales were very exercised by the loss of local employment that this caused.  I don’t know much about what happened up at Llanbedr but in Aberporth it has been an ongoing saga ever since the base was closed and handed over to the privatised company Qinetiq.  There was a lot of hoo haa at the time about the fact that that the management buyout team got the company for peanuts, more of this later.  This led to loss of hundreds of jobs and knocked on into the local economy.  In a separate though linked bit of economic development the Welsh Development Agency and the National Assembly For Wales set about using some of the £2.4 billion pounds of objective one EEC grant funding to set up a Science Park next to the the West Wales Airport which was bidding to become a full blown local airport with Civil Aviation Authority status (since gained).  The idea was that the Science Park would attract companies in the field of information and communications technology into West Wales and kick start economic regeneration along the lines of the Irish Celtic Tiger model which saw Eire become the second largest developers of software in the world.  High tech companies would be attracted by sweeteners such  low rents and rates and the close proximity to the airfield.  People were told that there would be replacement jobs and more for what had become an industrial wasteland.  The buildings were erected along with a swish new bit of roadway.  They looked fabulous and the WDA were sure they were onto a winner.  No high tech companies came, no jobs were created, and the local people watched a few million pounds wend down the Swanee.

The WDA and the National Assembly scratched their heads and worried about what was going to become of the Science Parc.  Meanwhile Qinetiq and the MOD were busy negotiating with the new owners of West Wales Airport to start using the newly extended runway to begin testing the Watchkeeper UAV for the British Army and so was hatched the plan to turn the science park into a brand new centre for excellence for Unmanned Aerial Systems Research and Development.  This seems to make absolute sense as there are many companies in the world who develop and  manufacture drone technology.  So with another fanfare of press releases saying that there would be 1000 new jobs created by the implementation of the new economic development opportunities afforded by drone technology companies from all over the world were invited to come and set up shop at the Science Park.  To date fewer than 50 actual jobs have been created by this initiative and most of the science park buildings remain unoccupied so after ten years the whole debacle looks like a wild west Wales ghost town with tumbleweed blowing around the brand spanking new empty roads.  There has been a lot of activity on the airport runway though as the MOD and Thales have carried out hundreds and hundreds of test flights of the Watchkeeper military surveillance drone.  Also Qinetiq and the MOD have organised a handful of drone trade fairs at the base but none of this activity has been of any major significance to the local economy whatsoever. I should also point out that the MOD applied for and got permission to extend the flight test area for the drones over a large area of Wales extending from Aberporth to Sennybridge.  The West Wales Airport has been granted permission for Unmanned Aerial Vehicle testing over large swathes of Welsh countryside.  Their was an, in my view, unsatisfactory public consultation carried out locally before the drone testing of the Watchkeeper actually began and people were told that the UAVs were not going to be solely used for military purposes, and in fact most of the drone development at Aberporth was going to be for civilian purposes.  Evidence of this is yet to be seen, in fact the only UAVs flying out of the airport are British Army Watchkeepers.

What is the current state of play vis a vis drone technology?

The whole issue of drone usage is controversial and is more and more coming onto the agenda of a variety of protest groups and other organisations around the world.  We all know that the governments of UK, USA and Israel are using drones as killing machines and for surveillance purposes on a continual basis.  So successful has the implementation of remote control technology been that we see decreasing levels of personnel employed by armed forces in what is euphemistically called the theatre of war.  This is however not exactly what all players in the game are happy with.  Politicians who are increasingly excercised by the tribulations of implementing austerity measures are very happy as drone warfare is cheaper than conventional warfare, and also makes life easier politically in that there are fewer personnel coming hope in black bags draped with flags.  However some sections of the armed services are worried about a number of issues arising from Pentagon driven so called advances in war fighting capabilities. In Britain the Raf have distinct qualms about the future developments in drone technology. The major issue is Autonomy.   The RAF say that “As UAVs are developed with increasing levels of automation it will reduce the requirement for operator training in the more traditional piloting skills of flying the aircraft, such as landing and takeoff, and focus the training more towards operating the payload.”

The MOD and in particular the RAF are discussing issues about the levels of automation they are comfortable with. They recognise that highly automated weaponry systems are unlikely to be able to apply judgement and pragmatism to “situations”. They are worried about legal and ethical considerations that occur when there are no human beings in the loop leading to loss of life or injury.  The future is one where outdated weaponry will give way to what they term “precision weaponry” and where the battle-spaces increasingly involve unmanned and cyber operations.

Autonomous weapon systems are capable of understanding higher level intent and direction, and perception of their environment, leading to the ability to take appropriate actions to bring about desired states. To be able to decide on a course of action, from a number of alternatives, without depending on human oversight and control. Although the overall activity of an autonomous unmanned aircraft will be predictable, individual actions may not be.

UAVs are used in scenarios which are highly distasteful and we all know that they are used for surveillance and targeting, and carrying out, weapon attacks inside the borders of countries such as Pakistan and Palestine. All of this is currently done with “man in the loop” systems but it will not be too long before drones are flying with what the RAF says is ” the ability to independently locate and attack mobile targets, with appropriate proportionality and discrimination.”

The RAF are worried about issues relating to the Geneva Convention with regard to autonomous UAV development and usage and they say that “compliance will become increasingly challenging as systems become more automated. In particular, if we wish to allow systems to make independent decisions without human intervention, some considerable work will be required to show how such systems will operate legally.”

Already there are automated weapons systems in use in Afghanistan, for example, the Phalanx and Counter-Rocket, Artillery and Mortar (C-RAM) systems used because there is deemed to be insufficient time for human response to counter incoming fire.

Future autonomous UAV systems will have to adhere to legal requirements and civilian airspace regulations. This will require political involvement in getting the necessary changes made. In my view it is absolutely vital that politicians understand the issues clearly and concisely because once made it will mean that there is going to be a lot of military and civilian hardware flying about in the airspace without any human interface whatsoever. The RAF say “As systems become increasingly automated, they will require decreasing human intervention between the issuing of mission-level orders and their execution.” and they go further saying “It would be only a small technical step to enable an unmanned aircraft to fire a weapon based solely on its own sensors, or shared information, and without recourse to higher, human authority.” They also discuss the timescale for the introduction of increased autonomy via Artificial Intelligence, saying ;- ” Estimates of when artificial intelligence will be achieved (as opposed to complex and clever automated systems) vary, but the consensus seems to lie between more than 5 years and less than 15 years.” Their words, not mine.

The RAF are clearly worried about the direction all this is going in and they say “As technology matures and new capabilities appear, policy-makers will need to be aware of the potential legal issues and take advice at a very early stage of any new system’s procurement cycle.” I believe this highlights a degree of paranoia on the part of the RAF vis a vis it’s own future role.

Not only are the RAF exercised about legal dilemmas. Ethics and morals related questions also are in their thoughts such as when, where and how automated and autonomous unmanned systems may be used. This applies not just to the use of drones of course but also to all other forms of weaponry in any environment.

Will all future wars be fought remotely with little or no loss of friendly or military personnel?

Will future conflicts be waged between increasingly complex unmanned systems?

In my view a problem that we face today is that big business accountants have control of governments and the most expensive resource used in public services is human beings. So autonomy offers massive savings in manpower and support for that manpower both before and after conflict occurs. As artificial Intelligence comes on board we are likely to see more complicated tasks arising that are beyond the capability of humans to deal with due to speed, complexity and information overload. No doubt some of you are probably suffering that now, but I only have a bit more to say before I will take questions, so bear with me.

The RAF and many others in the field are grappling with issues such as whether it is possible to develop AI that has the capability to focus on the unique (at the moment) ability that a human being has to bring empathy and morality to complex decision-making. The RAF say “To a robotic system, a school bus and a tank are the same – merely algorithms in a programme – and the engagement of a target is a singular action; the robot has no sense of ends, ways and means, no need to know why it is engaging a target. There is no recourse to human judgement in an engagement, no sense of a higher purpose on which to make decisions, and no ability to imagine (and therefore take responsibility for) repercussions of action taken.”

Why is there such acquiescence in the use of Drones by politicians from all the major political parties?

It is abundantly clear to me that politicians in Wales at all levels and in all parties are either happy to be part of a total militarised system or have no idea what is going on or are bereft of ways to get themselves out of their bankrupt policy morass vis a vis economic development and are therefore forced to clutch at the straws offered them by the military and their pals who run the transnational arms manufacturers.  In my view it is a combination of all three of these.  I can hear some of thinking what has all of this got to do with us in Wales?  Well Qinetiq, the NAFW and a number of transnational developers in drone and aerospace technology are engaged on a programme known as The ASTRAEA Project.  The ASTRAEA programme is jointly funded by industry and the public sector including the Welsh Assembly Government. Half of the funding for ASTRAEA is therefore being provided by public sector organisations – including the TSB and the regions – with the rest from a consortium of seven UK companies: AOS, BAE Systems, Cassidian, Cobham, QinetiQ, Rolls-Royce and Thales UK.

The TSB, Technology Strategy Board (ATory body set up David “two brains” Willets), is a business-led executive non-departmental public body, sponsored by the Department for Business, Innovation and Skills (BIS).

I know for a fact that the National Assembly For Wales has put at least £3 million into this project via it’s economic development commitee and I also know that this project was never discussed on the floor of the assembly in Cardiff.

ASTRAEA  stands for Autonomous Systems Technology Related Airborne Evaluation & Assessment and is a UK industry-led consortium focusing on the technologies, systems, facilities and procedures that will allow autonomous vehicles to operate safely and routinely in civil airspace over the United Kingdom.

The programme comprises two phases:

ASTRAEA I  –  2006 to 2008

Engaged with the regulators to understand the issues

Developed appropriate technology to TRL 3+

Culminated in an integrated SE demonstration of the art of the possible

Created an internationally recognised position for the UK

ASTRAEA II  –  2009 to 2013

Agreed a process with the regulator to establish suitability of various engineering solutions from the safety perspective

Created a process to identify and engage with potential UAS end users to establish more detailed and specific user requirements

Progressing solutions through live flight trials

Progress towards virtual certification

All of this research has clear military as well as civilian surveillance uses.  Don’t be fooled by UAV apologists who try to tell you that this research is benign, it most certainly is not.

How do we move forward to bring about an end to the development, testing and deployment of drones in Wales and beyond?

So we need to pose the following questions to evrybody out their be it our friends and neighbours, other members of the public, local journalists and media workers, and most of all our elected our politicians:-

Are they happy to allow Autonomous Robots to take on the responsibility of choosing which of us lives and which of us dies?

Can an autonomous robot be considered capable of waging ethical and legal warfare?

Can software and hardware accidents be distinguished from war crimes when robots go wrong?

Is it possible to establish clear policies on acceptable machine behavior?

How far out of the bottle is the technological genie?

Are we doomed to a Terminator style future?

Is it possible have a sensible debate about technological development anymore?

Do we really want Artificial Intelligence with a greater capacity to think than a human to be in any way involved in future theatres of War?

The RAF pose very important questions when they ask the following:-

– Do military planners and politicians understand the full implications of the systems they are currently tasking and those they hope to procure?

– In the current economic climate, who will decide the best balance between keeping existing equipment and personnel, or whether to give these up to fund new unmanned systems?

– Do we understand even the basic implications of such decisions for the associated defence lines of development?

– Crucially, do we have a strategic level of understanding as to how we will deliver the considerable number of changes that will need to be made to existing policy, concepts, doctrine, and force structures?

Finally, to conclude,  we must expect to see governments bringing in changes to the law of armed conflict in order to accommodate the use of autonomous UAS and we must shout our opposition to this from the rooftops.

So there we have it friends, there is a lot for us to consider when we look into the issue of drones. We have to view all of this in an holistic way. That is we must not just lumber along from one demo to the next thinking only about the impact of the current use of drone warfare, terrible though that is.

It is absolutely vital that we start to consider what future implications there are for the maintenance and development of basic human rights. We must also consider what the use of technology means in terms of social control measures such as we see the beginnings of in Gaza and elsewhere. I believe we are at, if not all ready passed, a dangerous turning point in the way we occupy this planet.

It is incumbent on us all to make a fuss about these issues if we want a planet where Human Rights are protected. If we don’t then we condemn the world to a dystopian future where all kinds of as yet un-thought of technology is used to maintain the position of a global elite above and beyond that of the majority of the people. The choice is ours, make a ruckus or bury our head in the sand and wait for the worst possible sci-fi future to engulf us all.  Drones Campaign Network Cymru is trying to ensure that as many people as we can reach are made aware of these issues, you can check us out on our Facebook page, we hope that we can be of help to people across Wales who want to try and influence a shift in policy decision making and eventually to bring about a drone free Wales.

Thanks for listening, I’m happy to take questions and take part in debate.

“BOOM”

“BOOM”
He stands, swaying gently from side to side
Weighing the heavy house brick in his hand
Staring intently at the curved shop window
It measures eleven feet high, fifteen feet wide
Standing behind the gleaming crystal screen
Dummies in their wool and worsted threads
Posed languidly, jaunty Trilbys on their heads
Seemingly inviting passers by to join them
In their blinding floodlit Burton’s wonderland
In his cider addled mind the choice is stark

Another long corrugated cardboard winter
In the grungy stairwell behind the slipper baths
Surrounded by paper litter and autumn leaves
With bottled Merrydown and rough scrumpy
To blot out the dankness in the piss stink dark
Or the next six months in Wormwood Scrubs
With three meals each day, dry bunk by night
Spinning yarns with long lost recidivist pals
A whole twenty five shillings saved each week
Pounding old lags sheets in laundry dolly tubs

Summoning strength he aims brick at glass
Falls backwards landing smack on his arse
Block hits window dead centre, corner first
Exploding casement showers shards at feet
Like wartime days when he could not be beat
On patrol shattered shops with each shell burst
Alarm loudly clanging, he steps into the display
Old bill handcuffing as they’re leading him away
“I am Harry Appletree, please take me to my room
I smashed that winder good, made a proper boom.”

Have you got all the footage you need….

Another of my posts from Facebook:-

As I watched the coverage on BBC news 24 of the students protesting in Parliament Square it seemed as if the police were deliberately allowing, indeed cajoling people making, acts of destruction and disfigurement in order that the media could have something to focus on.  The BBC journalists spent all day talking about the violence coming from the protestors and filming students smashing windows and trying to batter down doors on government buildings.  At the end of the day in Parliament Square the damage was superficial, some broken windows, a couple of phone boxes smashed up and daubed with paint and a few slogans sprayed onto buildings and monuments.  This was portrayed as if the country was facing the break down of law and order.  Not true, in fact the state want to be able introduce much tougher legislation in case there is more serious unrest against the ideological attack on the public services that the people of this country have spent generations nurturing and building.  The Coalition want to destroy common ownership and replace it with privatisation in as many areas of state provision as they possibly can get away with before 2015.  It was amazing to see the police stand back whilst “students” were allowed to aim a battering ram at the doors of the Supreme Court building for half an hour whilst Sky News and the BBC filmed it, eventually the police cleared the doorway and there was a policeman politely asking the camera operators “Have you got all the footage you need?”.  How revealing, the police had an agenda here as they did on a previous demo where a police van was deliberately allowed to be vandalised in order that pictures could be posted in the organs of the yellow pres, (Sun, Mail, Express, Star, Standard, Telegraph, Times, Standard etc) and used in TV programmes all across the globe to show what a lawless bunch of people the police are having to deal with.  Kettle the students in confined space, club a few of them, terrorise them with mounted officers weilding battons, and generally wind people up in order to provoke civil disobedience, this is the tactical approach of the police now that they have a sympathetic government in power that will not hesitate to give them whatever powers they need.  Political change is in the wind, young people are too clever to be hoodwinked by this approach, they are now getting organised in a way that was not envisaged so soon by the Tories.  In the new year when the ramifications of cuts start to hit people full force and mass redundancies start accruing, when VAT increases hit people in the pay-packet, when housing benefit cuts been people start losing their homes Cameron and his cronies want to ensure that they have the necessary powers to enforce order on the streets.  Of course as the events of Thursday 9th December 2010 show sometimes things get out of hand as the Chas and Camilla episode shows.  An unelected government bringing in policies made up on the hoof that were never put in front of the people is a recipe for disaster, either there will be an early election or there will be the implementation of a total police state, keep the cameras rolling, I need a little bit more footage…….

UN Report on Lethal Autonomous Robots (LARS)

United Nations A/HRC/23/47
General Assembly

Distr.: General
9 April 2013
Original: English

Human Rights Council
Twenty-third session
Agenda item 3
Promotion and protection of all human rights, civil,
political, economic, social and cultural

Report of the Special Rapporteur on extrajudicial,
summary or arbitrary executions, Christof Heyns
Summary
Lethal autonomous robotics (LARs) are weapon systems that, once activated, can select and engage targets without further human intervention. They raise far-reaching concerns about the protection of life during war and peace. This includes the question of the extent to which they can be programmed to comply with the requirements of international humanitarian law and the standards protecting life under international human rights law.  Beyond this, their deployment may be unacceptable because no adequate system of legal accountability can be devised, and because robots should not have the power of life and death over human beings. The Special Rapporteur recommends that States establish national moratoria on aspects of LARs, and calls for the establishment of a high level panel on LARs to articulate a policy for the international community on the issue.
United Nations A/HRC/23/47
General Assembly Distr.: General
9 April 2013
Original: English
A/HRC/23/47
2
Contents
Paragraphs Page
I. Introduction ………………………………………………………………………………………………. 1 3
II. Activities of the Special Rapporteur …………………………………………………………….. 2–25 3
A. Communications ……………………………………………………………………………….. 2–3 3
B. Visits ………………………………………………………………………………………………… 4–6 3
C. Press releases …………………………………………………………………………………….. 7–15 3
D. International and national meetings ………………………………………………………. 16–24 4
E. Intended future areas of research ………………………………………………………….. 25 5
III. Lethal autonomous robotics and the protection of life ……………………………………. 26–108 5
A. The emergence of LARs ……………………………………………………………………… 37–56 7
B. LARs and the decision to go to war or otherwise use force ………………………. 57–62 11
C. The use of LARs during armed conflict ………………………………………………… 63–74 12
D. Legal responsibility for LARs ……………………………………………………………… 75–81 14
E. The use of LARs by States outside armed conflict ………………………………….. 82–85 16
F. Implications for States without LARs ……………………………………………………. 86–88 16
G. Taking human decision-making out of the loop ……………………………………… 89–97 16
H. Other concerns …………………………………………………………………………………… 98–99 18
I. LARs and restrictive regimes on weapons ……………………………………………… 100–108 19
IV. Conclusions ………………………………………………………………………………………………. 109–112 20
V. Recommendations ……………………………………………………………………………………… 113–126 21
A. To the United Nations …………………………………………………………………………. 113–115 21
B. To regional and other inter-governmental organizations …………………………. 116–117 22
C. To States ………………………………………………………………………………………….. 118–121 22
D. To developers of robotic systems …………………………………………………………. 122 22
E. To NGOs, civil society and human rights groups and the ICRC ………………. 123–126 22
A/HRC/23/47
3
I. Introduction
1. The annual report of the Special Rapporteur on extrajudicial, summary and arbitrary executions, submitted to the Human Rights Council pursuant to its Resolution 17/5, focuses on lethal autonomous robotics and the protection of life.1 II. Activities of the Special Rapporteur
A. Communications
2. The present report covers communications sent by the Special Rapporteur between 16 March 2012 and 28 February 2013, and replies received between 1 May 2012 and 30 April 2013. The communications and responses from Governments are included in the following communications reports of special procedures: A/HRC/21/49; A/HRC/22/67 and A/HRC/23/51.
3. Observations on the communications sent and received during the reporting period are reflected in an addendum to the present report (A/HRC/23/47/Add.5).
B. Visits
4. The Special Rapporteur visited Turkey from 26 to 30 November 2012 and will visit Mexico from 22 April to 2 May 2013.
5. The Government of Mali has accepted the Special Rapporteur‟s visit requests and the Syrian Arab Republic views his proposal to visit the country positively. The Special Rapporteur thanks these Governments and encourages the Governments of Sri Lanka, the Republic of Madagascar and Pakistan to accept his pending requests for a visit.
6. Follow-up reports on missions undertaken by the previous mandate holder to Ecuador and Albania are contained in documents A/HRC/23/47/Add.3 and A/HRC/23/47/Add.4 respectively.
C. Press releases2
7. On 15 June 2012, the Special Rapporteur issued a joint statement with the Special Rapporteur on torture deploring the escalation of violence in the Syrian Arab Republic and called on all parties to renounce violence and lay down arms.
8. The Special Rapporteur issued several press releases with other mandate holders amongst others concerning aspects related to the right to life of human rights defenders in Honduras on 4 April 2012 and 1 October 2012; the Philippines, on 9 July 2012; and on 21 June 2012, he issued a press release to urge world governments, the international
1 The assistance of Tess Borden, Thompson Chengeta, Jiou Park and Jeff Dahlberg in writing this report is acknowledged with gratitude. The European University Institute is also thanked for hosting an expert consultation in February 2013, as well as the Global Justice Clinic, the Centre for Human Rights and Global Justice, and Professor Sarah Knuckey of New York University School of Law for preparing background materials and hosting an expert consultation in October 2012.
2 Press releases of the Special Rapporteur are available from http://www.ohchr.org/en/NewsEvents/Pages/NewsSearch.aspx?MID=SR_Summ_Executions.
A/HRC/23/47
4
community, journalists and media organizations to act decisively on the protection of the right to life of journalists and media freedom;
9. On 12 October 2012, a statement was sent jointly with other special rapporteurs concerning violence in Guatemala. The same day, the Special Rapporteur issued a joint statement regarding violence against a schoolchild in Pakistan.
10. On 22 October 2012, an open letter by special procedures mandate holders of the Human Rights Council was issued expressing concern at the planned adoption by the Congress of Colombia of a project to reform certain articles of the Political Constitution of Colombia, with regard to military criminal law.
11. On 15 November 2012, the Special Rapporteur, jointly with other mandate holders, called for an investigation into a death in custody in the Islamic Republic of Iran.
12. A joint statement was issued by all special procedures mandate holders on 23 November 2012 to express their dismay at the effect that the escalation of violence had on civilians in the Occupied Palestinian Territory and Israel.
13. On 28 February 2013, the Special Rapporteur together with other mandate holders called for an international inquiry into human rights violations in North Korea.
14. A number of press releases were issued specifically on death penalty cases concerning the following States: the United States of America, on 17 July 2012; Iraq, on 27 July 2012 and 30 August 2012; and the Gambia, on 28 August 2012.
15. Additional joint statements with other mandate holders on the death penalty were issued by the Special Rapporteur:
(a) The Islamic Republic of Iran: on 28 June 2012, concerning the execution of four individuals; on 12 October 2012, calling for a halt to executions; on 23 October 2012, regarding the execution of 10 individuals on drug-related crimes; and on 25 January 2013, urging the Iranian authorities to halt the execution of 5 Ahwazi activists;
(b) Saudi Arabia: on 11 January 2013, condemning the beheading of a domestic worker;
(c) Bangladesh: on 7 February 2013, expressing concern at a death sentence passed by the International Crimes Tribunal which failed to observe all the guarantees of a fair trial and due process.
D. International and national meetings
16. From 14 to 15 September 2012, the Special Rapporteur delivered a paper at the Pan-African Conference on the Safety of Journalists and the Issue of Impunity, held in Addis Ababa, Ethiopia.
17. On the occasion of the 52nd Ordinary Session of the African Commission on Human and Peoples‟ Rights on 9 October 2012, the Special Rapporteur delivered a statement on the cooperation between the United Nations and African Union special procedures mechanisms.
18. During the sixty-seventh session of the General Assembly, the Special Rapporteur was a panellist in the side-event on the theme “The Death Penalty and Human Rights”, organized by the Special Procedures Branch of the Office of the High Commissioner for Human Rights (OHCHR) in cooperation with the World Organisation Against Torture, Penal Reform International, the Center for Constitutional Rights and Human Rights Watch in New York on 24 October 2012.
A/HRC/23/47
5
19. On 25 October 2012, the Special Rapporteur participated in the weekly briefing entitled “Issue of the Moment: The Death Penalty” for the community of non-governmental organizations associated with the Department of Public Information in New York.
20. On 15 November 2012, the Special Rapporteur presented a lecture on “The Right to Life during Demonstrations” at a seminar organized by the South African Institute for Advanced Constitutional, Public, Human Rights and International Law at the Constitutional Court of South Africa in Johannesburg. On 22 and 23 November 2012, the Special Rapporteur was a panellist during the 2nd UN Inter-Agency meeting on the safety of journalists and the issue of impunity in Vienna, Austria.
21. The Special Rapporteur took part in an Expert Meeting in Geneva entitled “How Countries Abolished the Death Penalty”, organized by the International Commission against the Death Penalty on 5 February 2013, and delivered a presentation on the resumption of the death penalty.
22. On 22 February 2013, the Special Rapporteur participated in a High Level Policy Seminar organized by the European University Institute and Global Governance and Global Governance Programme on “Targeted Killing, Unmanned Aerial Vehicles and EU Policy”, held at the European University Institute in Florence, where he spoke on “Targeting by Drones: Protecting the Right to Life”.
23. On 19 March 2013, the Special Rapporteur presented a keynote address at a conference on “The Ethical, Strategic and Legal Implications of Drone Warfare”, organized by the Kroc Institute at the University of Notre Dame in Indiana, United States of America.
24. On 21 March 2013, the Special Rapporteur took part in the Pugwash Workshop at the University of Birmingham, United Kingdom, where he spoke on lethal autonomous robotics.
E. Intended future areas of research
25. The Special Rapporteur will present a report on unmanned combat aerial vehicles (UCAVs) to the General Assembly in 2013.
III. Lethal autonomous robotics and the protection of life
26. For societies with access to it, modern technology allows increasing distance to be put between weapons users and the lethal force they project. For example, UCAVs, commonly known as drones, enable those who control lethal force not to be physically present when it is deployed, but rather to activate it while sitting behind computers in faraway places, and stay out of the line of fire.
27. Lethal autonomous robotics (LARs), if added to the arsenals of States, would add a new dimension to this distancing, in that targeting decisions could be taken by the robots themselves. In addition to being physically removed from the kinetic action, humans would also become more detached from decisions to kill – and their execution.
28. The robotics revolution has been described as the next major revolution in military affairs, on par with the introduction of gunpowder and nuclear bombs.3 But in an important respect LARs are different from these earlier revolutions: their deployment would entail not merely an upgrade of the kinds of weapons used, but also a change in the identity of those
3 Peter Singer, Wired for War (Penguin Group (USA) Incorporated, 2009), p. 179 and further,notably p. 203.
A/HRC/23/47
6
who use them. With the contemplation of LARs, the distinction between weapons and warriors risks becoming blurred, as the former would take autonomous decisions about their own use.
29. Official statements from Governments with the ability to produce LARs indicate that their use during armed conflict or elsewhere is not currently envisioned.4 While this may be so, it should be recalled that aeroplanes and drones were first used in armed conflict for surveillance purposes only, and offensive use was ruled out because of the anticipated adverse consequences.5 Subsequent experience shows that when technology that provides a perceived advantage over an adversary is available, initial intentions are often cast aside. Likewise, military technology is easily transferred into the civilian sphere. If the international legal framework has to be reinforced against the pressures of the future, this must be done while it is still possible.
30. One of the most difficult issues that the legal, moral and religious codes of the world have grappled with is the killing of one human being by another. The prospect of a future in which fully autonomous robots could exercise the power of life and death over human beings raises a host of additional concerns. As will be argued in what follows, the introduction of such powerful yet controversial new weapons systems has the potential to pose new threats to the right to life. It could also create serious international division and weaken the role and rule of international law – and in the process undermine the international security system.6 The advent of LARs requires all involved – States, international organizations, and international and national civil societies – to consider the full implications of embarking on this road.
31. Some argue that robots could never meet the requirements of international humanitarian law (IHL) or international human rights law (IHRL), and that, even if they could, as a matter of principle robots should not be granted the power to decide who should live and die. These critics call for a blanket ban on their development, production and use.7 To others, such technological advances – if kept within proper bounds – represent legitimate military advances, which could in some respects even help to make armed conflict more humane and save lives on all sides.8 According to this argument, to reject this technology altogether could amount to not properly protecting life.
32. However, there is wide acceptance that caution and some form of control of States‟ use of this technology are needed, over and above the general standards already posed by international law. Commentators agree that an international discussion is needed to consider the appropriate approach to LARs.
4 US Department of Defense, Unmanned Systems Integrated Road Map FY2011-2036, p. 50, available from http://publicintelligence.net/dod-unmanned-systems-integrated-roadmap-fy2011-2036
5 See http://www.usaww1.com/World_War_1_Fighter_Planes.php4
6 Nils Melzer, “Human rights implications of the usage of drones and unmanned robots in warfare” Study for the European Parliament‟s Subcommittee on Human Rights (2013), available from http://www.europarl.europa.eu/committees/en/studies/html, p. 5 (forthcoming).
7 Human Rights Watch, Losing Humanity: The Case Against Killer Robots (2012), p. 2, available from http://www.hrw.org/reports/2012/11/19/losing-humanity-0. See in response Michael Schmitt “Autonomous Weapons Systems and International Humanitarian Law: A Reply to the Critics” Harvard International Security Journal (forthcoming 2013), available from http://harvardnsj.org/wp-content/uploads/2013/02/Schmitt-Autonomous-Weapon-Systems-and-IHL-Final.pdf). The International Committee on Robot Arms Control (ICRAC) was formed to promote such a ban. See http://icrac.net
8 Ronald Arkin, Governing Lethal Behaviour in Autonomous Robots (CRC Press, 2009); Kenneth Anderson and Matthew Waxman, “Law and ethics for robot soldiers”, Policy Review, No. 176 (2012), available from http://www.hoover.org/publications/policy-review/article/135336.
A/HRC/23/47
7
33. As with any technology that revolutionizes the use of lethal force, little may be known about the potential risks of the technology before it is developed, which makes formulating an appropriate response difficult; but afterwards the availability of its systems and the power of vested interests may preclude efforts at appropriate control.9 This is further complicated by the arms race that could ensue when only certain actors have weapons technology. The current moment may be the best we will have to address these concerns. In contrast to other revolutions in military affairs, where serious reflection mostly began after the emergence of new methods of warfare, there is now an opportunity collectively to pause, and to engage with the risks posed by LARs in a proactive way. This report is a call for pause, to allow serious and meaningful international engagement with this issue.
34. One of the reasons for the urgency of this examination is that current assessments of the future role of LARs will affect the level of investment of financial, human and other resources in the development of this technology over the next several years. Current assessments – or the lack thereof – thus risk to some extent becoming self-fulfilling prophesies.
35. The previous Special Rapporteur examined LARs in a report in 2010,10 calling inter alia for the convening of an expert group to consider robotic technology and compliance with international human rights and humanitarian law.11 The present report repeats and strengthens that proposal and calls on States to impose national moratoria on certain activities related to LARs.
36. As with UCAVs and targeted killing, LARs raise concerns for the protection of life under the framework of IHRL as well as IHL. The Special Rapporteur recalls the supremacy and non-derogability of the right to life under both treaty and customary international law.12 Arbitrary deprivation of life is unlawful in peacetime and in armed conflict.
A. The emergence of LARs
1. Definitions
37. While definitions of the key terms may differ, the following exposition provides a starting point.13
38. According to a widely used definition (endorsed inter alia by both the United States Department of Defense and Human Rights Watch14), the term LARs refers to robotic weapon systems that, once activated, can select and engage targets without further
9 David Collingridge, The Social Control of Technology (Frances Pinter, 1980).
10 A/65/321.
11 A/65/321, pp. 10-22.
12 International Covenant on Civil and Political Rights, art. 6, enshrining the right to life, and art. 4 (2) on its non-derogability.
13 Arkin (see note 8 above), p. 7; Noel Sharkey AutomatingWarfare: lessons learned from the drones, p. 2, available from http://www.alfredoroma.it/wp-content/uploads/2012/05/Automated-warfare-Noel-Sharkey.pdf; Patrick Lin et al, Autonomous Military Robotics: Risk, Ethics, and Design (San Luis Obispo, California Polytechnic State University, 2008) p. 4, available from http://ethics.calpoly.edu/ONR_report.pdf
14 US Department of Defense Directive, “Autonomy in Weapons Systems”, Number 3000.09 of 21 November 2012, Glossary Part II. See also United Kingdom Ministry of Defence “The UK Approach to Unmanned Aircraft Systems” paras. 202-203, available from https://www.gov.uk/government/publications/jdn-2-11-the-uk-approach-to-unmanned-aircraft-systems; see also, Human Rights Watch (see note 7 above), p. 2.
A/HRC/23/47
8
intervention by a human operator. The important element is that the robot has an autonomous “choice” regarding selection of a target and the use of lethal force.
39. Robots are often described as machines that are built upon the sense-think-act paradigm: they have sensors that give them a degree of situational awareness; processors or artificial intelligence that “decides” how to respond to a given stimulus; and effectors that carry out those “decisions”.15 The measure of autonomy that processors give to robots should be seen as a continuum with significant human involvement on one side, as with UCAVs where there is “a human in the loop”, and full autonomy on the other, as with LARs where human beings are “out of the loop”.
40. Under the currently envisaged scenario, humans will at least remain part of what may be called the “wider loop”: they will programme the ultimate goals into the robotic systems and decide to activate and, if necessary, deactivate them, while autonomous weapons will translate those goals into tasks and execute them without requiring further human intervention.
41. Supervised autonomy means that there is a “human on the loop” (as opposed to “in” or “out”), who monitors and can override the robot‟s decisions. However, the power to override may in reality be limited because the decision-making processes of robots are often measured in nanoseconds and the informational basis of those decisions may not be practically accessible to the supervisor. In such circumstances humans are de facto out of the loop and the machines thus effectively constitute LARs.
42. “Autonomous” needs to be distinguished from “automatic” or “automated.” Automatic systems, such as household appliances, operate within a structured and predictable environment. Autonomous systems can function in an open environment, under unstructured and dynamic circumstances. As such their actions (like those of humans) may ultimately be unpredictable, especially in situations as chaotic as armed conflict, and even more so when they interact with other autonomous systems.
43. The terms “autonomy” or “autonomous”, as used in the context of robots, can be misleading. They do not mean anything akin to “free will” or “moral agency” as used to describe human decision-making. Moreover, while the relevant technology is developing at an exponential rate, and full autonomy is bound to mean less human involvement in 10 years‟ time compared to today, sentient robots, or strong artificial intelligence are not currently in the picture.16
2. Current technology
44. Technology may in some respects be less advanced than is suggested by popular culture, which often assigns human-like attributes to robots and could lure the international community into misplaced trust in its abilities. However, it should also be recalled that in certain respects technology far exceeds human ability. Technology is developing exponentially, and it is impossible to predict the future confidently. As a result, it is almost impossible to determine how close we are to fully autonomous robots that are ready for use.
45. While much of their development is shrouded in secrecy, robots with full lethal autonomy have not yet been deployed. However, robotic systems with various degrees of autonomy and lethality are currently in use, including the following:
• The US Phalanx system for Aegis-class cruisers automatically detects, tracks and engages anti-air warfare threats such as anti-ship missiles and aircraft.17
15 Singer (see note 3 above), p. 67.
16 The same applies to “the Singularity”, Singer (see note 3 above), p. 101.
17 See http://usmilitary.about.com/library/milinfo/navyfacts/blphalanx.htm
A/HRC/23/47
9
• The US Counter Rocket, Artillery and Mortar (C-RAM) system can automatically destroy incoming artillery, rockets and mortar rounds.18
• Israel‟s Harpy is a “Fire-and-Forget” autonomous weapon system designed to detect, attack and destroy radar emitters.19
• The United Kingdom Taranis jet-propelled combat drone prototype can autonomously search, identify and locate enemies but can only engage with a target when authorized by mission command. It can also defend itself against enemy aircraft.20
• The Northrop Grumman X-47B is a fighter-size drone prototype commissioned by the US Navy to demonstrate autonomous launch and landing capability on aircraft carriers and navigate autonomously.21
• The Samsung Techwin surveillance and security guard robots, deployed in the demilitarized zone between North and South Korea, detect targets through infrared sensors. They are currently operated by humans but have an “automatic mode”.22
46. Military documents of a number of States describe air, ground and marine robotic weapons development programmes at various stages of autonomy. Large amounts of money are allocated for their development.23
47. It seems clear that if introduced, LARs will not, at least initially, entirely replace human soldiers, but that they will have discretely assigned tasks suitable to their specific capabilities. Their most likely use during armed conflict is in some form of collaboration with humans,24 although they would still be autonomous in their own functions. The question should therefore be asked to what extent the existing legal framework is sufficient to regulate this scenario, as well as the scenario whereby LARs are deployed without any human counterpart. Based on current experiences of UCAVs, there is reason to believe that States will inter alia seek to use LARs for targeting killing.
48. The nature of robotic development generally makes it a difficult subject of regulation, especially in the area of weapons control. Bright lines are difficult to find. Robotic development is incremental in nature. Furthermore, there is significant continuity between military and non-military technologies.25 The same robotic platforms can have civilian as well as military applications, and can be deployed for non-lethal purposes (e.g. to defuse improvised explosive devices) or be equipped with lethal capability (i.e. LARs). Moreover, LARs typically have a composite nature and are combinations of underlying technologies with multiple purposes.
49. The importance of the free pursuit of scientific study is a powerful disincentive to regulate research and development in this area. Yet “technology creep” in this area may
18 See http://www.dtic.mil/cgi-bin/GetTRDoc?AD=ADA557876
19 See http://www.israeli-weapons.com/weapons/aircraft/uav/harpy/harpy.html
20 See http://www.baesystems.com/product/BAES_020273/taranis
21 See http://www.as.northropgrumman.com/products/nucasx47b/assets/X-47B_Navy_UCAS_FactSheet.pdf
22 See http://singularityhub.com/2010/07/25/armed-robots-deployed-by-south-korea-in-demilitarized-zone-on-trial-basis
23 United States Air Force, “UAS Flight Plan 2009-2047” (Washington, D.C., 2009) p. 41, available from http://www.scribd.com/doc/17312080/United-States-Air-Force-Unmanned-Aircraft-Systems-Flight-Plan-20092047-Unclassified
24 Ronald Arkin “Governing Lethal Behaviour: Embedding Ethics in a Hybrid Deliberative/Reactive Robot Architecture” Technical Report GIT-GVU-07-11 p. 5, available from http://www.cc.gatech.edu/ai/robot-lab/online-publications/formalizationv35.pdf
25 Anderson and Waxman (see note 8 above), pp. 2 and 13 and Singer (see note 3 above), p. 379.
A/HRC/23/47
10
over time and almost unnoticeably result in a situation which presents grave dangers to core human values and to the international security system. It is thus essential for the international community to take stock of the current state of affairs, and to establish a responsible process to address the situation and where necessary regulate the technology as it develops.
3. Drivers of and impediments to the development of LARs
50. Some of the reasons to expect continuous pressures to develop LARs, as well as the impediments to this momentum, also apply to the development of other unmanned systems more generally. They offer huge military and other advantages to those using them and are part of the broader automization of warfare and of the world in general.
51. Unmanned systems offer higher force projection (preserving the lives of one‟s own soldiers) and force multiplication (allowing fewer personnel to do more). They are capable of enlarging the battlefield, penetrating more easily behind enemy lines, and saving on human and financial resources. Unmanned systems can stay on station much longer than individuals and withstand other impediments such as G-forces. They can enhance the quality of life of soldiers of the user party: unmanned systems, especially robots, are increasingly developed to do the so-called dirty, dull and dangerous work.26
52. Robots may in some respects serve humanitarian purposes. While the current emergence of unmanned systems may be related to the desire on the part of States not to become entangled in the complexities of capture, future generations of robots may be able to employ less lethal force, and thus cause fewer unnecessary deaths. Technology can offer creative alternatives to lethality, for instance by immobilizing or disarming the target.27 Robots can be programmed to leave a digital trail, which potentially allows better scrutiny of their actions than is often the case with soldiers and could therefore in that sense enhance accountability.
53. The progression from remote controlled systems to LARs, for its part, is driven by a number of other considerations.28 Perhaps foremost is the fact that, given the increased pace of warfare, humans have in some respects become the weakest link in the military arsenal and are thus being taken out of the decision-making loop. The reaction time of autonomous systems far exceeds that of human beings, especially if the speed of remote-controlled systems is further slowed down through the inevitable time-lag of global communications. States also have incentives to develop LARs to enable them to continue with operations even if communication links have been broken off behind enemy lines.
54. LARs will not be susceptible to some of the human shortcomings that may undermine the protection of life. Typically they would not act out of revenge, panic, anger, spite, prejudice or fear. Moreover, unless specifically programmed to do so, robots would not cause intentional suffering on civilian populations, for example through torture. Robots also do not rape.
55. Yet robots have limitations in other respects as compared to humans. Armed conflict and IHL often require human judgement, common sense, appreciation of the larger picture, understanding of the intentions behind people‟s actions, and understanding of values and anticipation of the direction in which events are unfolding. Decisions over life and death in armed conflict may require compassion and intuition. Humans – while they are fallible – at least might possess these qualities, whereas robots definitely do not. While
26 Gary Marchant et al, “International governance of autonomous military robots”, Columbia Science and Technology Law Review, Volume XII (2011) p. 275.
27 Singer (see note 3 above), p. 83.
28 Arkin (see note 8 above), xii.
A/HRC/23/47
11
robots are especially effective at dealing with quantitative issues, they have limited abilities to make the qualitative assessments that are often called for when dealing with human life. Machine calculations are rendered difficult by some of the contradictions often underlying battlefield choices. A further concern relates to the ability of robots to distinguish legal from illegal orders.
56. While LARs may thus in some ways be able to make certain assessments more accurately and faster than humans, they are in other ways more limited, often because they have restricted abilities to interpret context and to make value-based calculations.
B. LARs and the decision to go to war or otherwise use force
57. During the larger part of the last two centuries, international law was developed to constrain armed conflict and the use of force during law enforcement operations, to make it an option of last resort. However, there are also built-in constraints that humans have against going to war or otherwise using force which continue to play an important (if often not decisive) role in safeguarding lives and international security. Chief among these are unique human traits such as our aversion to getting killed, losing loved ones, or having to kill other people.29 The physical and psychological distance from the actual use of force potentially introduced by LARs can lessen all three concerns and even render them unnoticeable to those on the side of the State deploying LARs.30 Military commanders for example may therefore more readily deploy LARs than real human soldiers.
58. This ease could potentially affect political decisions. Due to the low or lowered human costs of armed conflict to States with LARs in their arsenals, the national public may over time become increasingly disengaged and leave the decision to use force as a largely financial or diplomatic question for the State, leading to the “normalization” of armed conflict.31 LARs may thus lower the threshold for States for going to war or otherwise using lethal force, resulting in armed conflict no longer being a measure of last resort32 According to the report of the Secretary-General on the role of science and technology in the context of international security and disarmament, “…the increased capability of autonomous vehicles opens up the potential for acts of warfare to be conducted by nations without the constraint of their people‟s response to loss of human life.”33 Presenting the use of unmanned systems as a less costly alternative to deploying “boots on the ground” may thus in many cases be a false dichotomy. If there is not sufficient support for a ground invasion, the true alternative to using unmanned systems may be not to use force at all.
59. Some have argued that if the above reasoning is taken to its logical conclusion, States should not attempt to develop any military technology that reduces the brutality of armed conflict or lowers overall deaths through greater accuracy.34 Drones and high-altitude airstrikes using smart bombs should then equally be viewed as problematic because
29 A/65/321, para. 44; John Mueller “The Iraq Syndrome”, Foreign Affairs, Vol. 84, No. 6, p. 44 (November/December 2005).
30 According to military experts, it generally becomes easier to take life as the distance between the actor and the target increases. See David Grossman On Killing: The Psychological Cost of Learning to Kill in War and Society (Back Bay Books, 1996).
31 Armin Krishnan Killer robots: Legality and Ethicality of Autonomous Weapons (Ashgate, 2009) p. 150
32 Singer (see note 3 above), p. 323; Peter Asaro “How Just Could a Robot War Be?” in P. Brey et al (eds.) Current Issues in Computing And Philosophy (2008), p. 7.
33 A/53/202, para. 98.
34 Asaro (see note 32 above), pp. 7-9. Discussed by Patrick Lin et al “Robots in War: Issues of Risk and Ethics” in R. Capurro & M. Nagenborg (eds.) Ethics and Robotics (2009) p. 57.
A/HRC/23/47
12
they also lower casualty rates for the side that uses them (and in some cases also for the other side), thereby removing political constraints on States to resort to military action.35
60. This argument does not withstand closer scrutiny. While it is desirable for States to reduce casualties in armed conflict, it becomes a question whether one can still talk about “war” – as opposed to one-sided killing – where one party carries no existential risk, and bears no cost beyond the economic. There is a qualitative difference between reducing the risk that armed conflict poses to those who participate in it, and the situation where one side is no longer a “participant” in armed conflict inasmuch as its combatants are not exposed to any danger.36 LARs seem to take problems that are present with drones and high-altitude airstrikes to their factual and legal extreme.
61. Even if it were correct to assume that if LARs were used there would sometimes be fewer casualties per armed conflict, the total number of casualties in aggregate could still be higher.
62. Most pertinently, the increased precision and ability to strike anywhere in the world, even where no communication lines exist, suggests that LARs will be very attractive to those wishing to perform targeted killing. The breaches of State sovereignty – in addition to possible breaches of IHL and IHRL – often associated with targeted killing programmes risk making the world and the protection of life less secure.
C. The use of LARs during armed conflict
63. A further question is whether LARs will be capable of complying with the requirements of IHL. To the extent that the answer is negative, they should be prohibited weapons. However, according to proponents of LARs this does not mean that LARs are required never to make a mistake – the yardstick should be the conduct of human beings who would otherwise be taking the decisions, which is not always a very high standard.37
64. Some experts have argued that robots can in some respects be made to comply even better with IHL requirements than human beings.38 Roboticist Ronald Arkin has for example proposed ways of building an “ethical governor” into military robots to ensure that they satisfy those requirements.39
65. A consideration of a different kind is that if it is technically possible to programme LARs to comply better with IHL than the human alternatives, there could in fact be an obligation to use them40 – in the same way that some human rights groups have argued that where available, “smart” bombs, rather than less discriminating ones, should be deployed.
66. Of specific importance in this context are the IHL rules of distinction and proportionality. The rule of distinction seeks to minimize the impact of armed conflict on civilians, by prohibiting targeting of civilians and indiscriminate attacks.41 In situations
35 Anderson and Waxman (see note 8 above), p. 12.
36 According to some commentators, war requires some willingness to accept reciprocal or mutual risk, involving some degree of sacrifice. See Paul Kahn “The Paradox of Riskless Warfare” Philosophy and Public Policy Vol. 22 (2002) and “War and Sacrifice in Kosovo” (1999), available from http://www-personal.umich.edu/~elias/Courses/War/kosovo.htm
37 Lin (see note 34 above), p. 50.
38 Marchant (see note 26 above), p. 280; Singer, (see note 3 above), p. 398.
39 Arkin (see note 8 above), p. 127.
40 Jonathan Herbach “Into the Caves of Steel: Precaution, Cognition and Robotic Weapons Systems Under the International Law of Armed Conflict” Amsterdam Law Forum Vol. 4 (2012), p. 14.
41 Protocol I additional to the Geneva Conventions, 1977, arts. 51 and 57.
A/HRC/23/47
13
where LARs cannot reliably distinguish between combatants or other belligerents and civilians, their use will be unlawful.
67. There are several factors that will likely impede the ability of LARs to operate according to these rules in this regard, including the technological inadequacy of existing sensors,42 a robot‟s inability to understand context, and the difficulty of applying of IHL language in defining non-combatant status in practice, which must be translated into a computer programme.43 It would be difficult for robots to establish, for example, whether someone is wounded and hors de combat, and also whether soldiers are in the process of surrendering.
68. The current proliferation of asymmetric warfare and non-international armed conflicts, also in urban environments, presents a significant barrier to the capabilities of LARs to distinguish civilians from otherwise lawful targets. This is especially so where complicated assessments such as “direct participation in hostilities” have to be made. Experts have noted that for counter-insurgency and unconventional warfare, in which combatants are often only identifiable through the interpretation of conduct, the inability of LARs to interpret intentions and emotions will be a significant obstacle to compliance with the rule of distinction.44
69. Yet humans are not necessarily superior to machines in their ability to distinguish. In some contexts technology can offer increased precision. For example, a soldier who is confronted with a situation where it is not clear whether an unknown person is a combatant or a civilian may out of the instinct of survival shoot immediately, whereas a robot may utilize different tactics to go closer and, only when fired upon, return fire. Robots can thus act “conservatively”45 and “can shoot second.”46 Moreover, in some cases the powerful sensors and processing powers of LARs can potentially lift the “fog of war” for human soldiers and prevent the kinds of mistakes that often lead to atrocities during armed conflict, and thus save lives.47
70. The rule of proportionality requires that the expected harm to civilians be measured, prior to the attack, against the anticipated military advantage to be gained from the operation.48 This rule, described as “one of the most complex rules of international humanitarian law,”49 is largely dependent on subjective estimates of value and context-specificity.
71. Whether an attack complies with the rule of proportionality needs to be assessed on a case-by-case basis, depending on the specific context and considering the totality of the circumstances.50 The value of a target, which determines the level of permissible collateral damage, is constantly changing and depends on the moment in the conflict. Concerns have been expressed that the open-endedness of the rule of proportionality combined with the complexity of circumstances may result in undesired and unexpected behaviour by LARs, with deadly consequences.51 The inability to “frame” and contextualize the environment
42 Noel Sharkey “Grounds for Discrimination: Autonomous Robot Weapons” RUSI Defence Systems (Oct 2008) pp. 88-89, available from http://rusi.org/downloads/assets/23sharkey.pdf
43 Peter Asaro “On Banning Autonomous Weapon Systems: Human Rights, Automation, and the Dehumanisation of Lethal Decision-making” p. 94, International Review of the Red Cross (forthcoming 2013) p. 11.
44 Human Rights Watch (see note 7 above), p. 31.
45 Marchant (see note 26 above), p. 280.
46 Singer (see note 3 above), p. 398.
47 Ibid.
48 Protocol I additional to the Geneva Conventions, 1977, art. 51 (5) (b).
49 Human Rights Watch (see note 7 above), p. 32.
50 Lin (see note 34 above), p. 57.
51 Noel Sharkey, “Automated Killers and the Computing Profession” Computer, Vol. 40 (2007), p. 122.
A/HRC/23/47
14
may result in a LAR deciding to launch an attack based not merely on incomplete but also on flawed understandings of the circumstances.52 It should be recognized, however, that this happens to humans as well.
72. Proportionality is widely understood to involve distinctively human judgement. The prevailing legal interpretations of the rule explicitly rely on notions such as “common sense”, “good faith” and the “reasonable military commander standard.”53 It remains to be seen to what extent these concepts can be translated into computer programmes, now or in the future.
73. Additionally, proportionality assessments often involve qualitative rather than quantitative judgements.54
74. In view of the above, the question arises as to whether LARs are in all cases likely (on the one hand) or never (on the other) to meet this set of cumulative standard. The answer is probably less absolute, in that they may in some cases meet them (e.g. in the case of a weapons system that is set to only return fire and that is used on a traditional battlefield) but in other cases not (e.g. where a civilian with a large piece of metal in his hands must be distinguished from a combatant in plain clothes). Would it then be possible to categorize the different situations, to allow some to be prohibited and others to be permitted? Some experts argue that certain analyses such as proportionality would at least initially have to be made by commanders, while other aspects could be left to LARs.55
D. Legal responsibility for LARs
75. Individual and State responsibility is fundamental to ensure accountability for violations of international human rights and international humanitarian law. Without the promise of accountability, deterrence and prevention are reduced, resulting in lower protection of civilians and potential victims of war crimes.56
76. Robots have no moral agency and as a result cannot be held responsible in any recognizable way if they cause deprivation of life that would normally require accountability if humans had made the decisions. Who, then, is to bear the responsibility?
77. The composite nature of LAR technology and the many levels likely to be involved in decisions about deployment result in a potential accountability gap or vacuum. Candidates for legal responsibility include the software programmers, those who build or sell hardware, military commanders, subordinates who deploy these systems and political leaders.
78. Traditionally, criminal responsibility would first be assigned within military ranks. Command responsibility should be considered as a possible solution for accountability for
52 Krishnan, (see note 31 above), pp. 98-99.
53 Tonya Hagmaier et al, “Air force operations and the law: A guide for air, space and cyber forces” p. 21, available from http://www.afjag.af.mil/shared/media/document/AFD-100510-059.pdf; Andru Wall “Legal and Ethical Lessons of NATO‟s Kosovo Campaign” p. xxiii, available from http://www.au.af.mil/au/awc/awcgate/navy/kosovo_legal.pdf
54 Markus Wagner “The Dehumanization of International Humanitarian Law: Legal, Ethical, and Political Implications of Autonomous Weapon Systems” (2012), available from http://robots.law.miami.edu/wp-content/uploads/2012/01/Wagner_Dehumanization_of_international_humanitarian_law.pdf note 96 and accompanying text.
55 Benjamin Kastan “Autonomous Weapons Systems: A Coming Legal „Singularity‟?” University of Illinois Journal of Law, Technology and Policy (forthcoming 2013), p. 18 and further, available from http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2037808
56 Human Rights Watch (see note 7 above), pp. 42-45.
A/HRC/23/47
15
LAR violations.57 Since a commander can be held accountable for an autonomous human subordinate, holding a commander accountable for an autonomous robot subordinate may appear analogous. Yet traditional command responsibility is only implicated when the commander “knew or should have known that the individual planned to commit a crime yet he or she failed to take action to prevent it or did not punish the perpetrator after the fact.”58 It will be important to establish, inter alia, whether military commanders will be in a position to understand the complex programming of LARs sufficiently well to warrant criminal liability.
79. It has been proposed that responsibility for civil damages at least should be assigned to the programmer and the manufacturers, by utilizing a scheme similar to strict product liability. Yet national product liability laws remain largely untested in regard to robotics.59 The manufacturing of a LAR will invariably involve a vast number of people, and no single person will be likely to understand the complex interactions between the constituent components of LARs.60 It is also questionable whether putting the onus of bringing civil suits on victims is equitable, as they would have to bring suit while based in a foreign country, and would often lack the resources.
80. The question of legal responsibility could be an overriding issue. If each of the possible candidates for responsibility identified above is ultimately inappropriate or impractical, a responsibility vacuum will emerge, granting impunity for all LAR use. If the nature of a weapon renders responsibility for its consequences impossible, its use should be considered unethical and unlawful as an abhorrent weapon.61
81. A number of novel ways to establish legal accountability could be considered. One of the conditions that could be imposed for the use of LARs is that responsibility is assigned in advance.62 Due to the fact that technology potentially enables more precise monitoring and reconstruction of what occurs during lethal operations, a further condition for their use could be the installation of such recording devices, and the mandatory ex post facto review of all footage in cases of lethal use, regardless of the status of the individual killed.63 A system of “splitting” responsibility between the potential candidates could also be considered.64 In addition, amendments to the rules regarding command responsibility may be needed to cover the use of LARs. In general, a stronger emphasis on State as opposed to individual responsibility may be called for, except in respect of its use by non-state actors.
57 Rome Statute of the ICC, art. 28; Heather Roff “Killing in War: Responsibility, Liability and Lethal Autonomous Robots” p. 14, available from http://www.academia.edu/2606840/Killing_in_War_Responsibility_Liability_and_Lethal_Autonomous_Robots
58 Protocol I additional to the Geneva Conventions, 1977, arts. 86 (2) and 87.
59 Patrick Lin “Introduction to Robot Ethics” in Patrick Lin et al (eds.) Robot Ethics: The ethical and Social Implications of Robotics (MIT Press, 2012), p. 8.
60 Wendell Wallach “From Robots to Techno Sapiens: Ethics, Law and Public Policy in the Development of Robotics and Neurotechnologies” Law, Innovation and Technology Vol. 3 (2011) p. 194.
61 Gianmarco Verugio and Keith Abney “Roboethics: The Applied Ethics for a New Science” in Lin, (see note 59 above), p. 114; Robert Sparrow “Killer Robots” Journal of Applied Philosophy Vol. 24, No. 1 (2007).
62 See Ronald Arkin “The Robot didn‟t do it” Position Paper for the Workshop on Anticipatory Ethics, Responsibility and Artificial Agents p. 1, available from http://www.cc.gatech.edu/ai/robot-lab/publications.html
63 Marchant (see note 26 above), p. 7.
64 Krishnan (see note 31 above), 105.
A/HRC/23/47
16
E. The use of LARs by States outside armed conflict
82. The experience with UCAVs has shown that this type of military technology finds its way with ease into situations outside recognized battlefields.
83. One manifestation of this, whereby ideas of the battlefield are expanded beyond IHL contexts, is the situation in which perceived terrorists are targeted wherever they happen to be found in the world, including in territories where an armed conflict may not exist and IHRL is the applicable legal framework. The danger here is that the world is seen as a single, large and perpetual battlefield and force is used without meeting the threshold requirements. LARs could aggravate these problems.
84. On the domestic front, LARs could be used by States to suppress domestic enemies and to terrorize the population at large, suppress demonstrations and fight “wars” against drugs. It has been said that robots do not question their commanders or stage coups d‟état.65
85. The possibility of LAR usage in a domestic law enforcement situation creates particular risks of arbitrary deprivation of life, because of the difficulty LARs are bound to have in meeting the stricter requirements posed by IHRL.
F. Implications for States without LARs
86. Phrases such as “riskless war” and “wars without casualties” are often used in the context of LARs. This seems to purport that only the lives of those with the technology count, which suggests an underlying concern with the deployment of this technology, namely a disregard for those without it. LARs present the ultimate asymmetrical situation, where deadly robots may in some cases be pitted against people on foot. LARs are likely – at least initially – to shift the risk of armed conflict to the belligerents and civilians of the opposing side.
87. The use of overwhelming force has proven to have counterproductive results – e.g. in the context of demonstrations, where psychologists warn that it may elicit escalated counter force66 In situations of hostilities, the unavailability of a legitimate human target of the LAR user State on the ground may result in attacks on its civilians as the “best available” targets and the use of LARs could thus possibly encourage retaliation, reprisals and terrorism.67
88. The advantage that States with LARs would have over others is not necessarily permanent. There is likely to be proliferation of such systems, not only to those to which the first user States transfer and sell them. Other States will likely develop their own LAR technology, with inter alia varying degrees of IHL-compliant programming, and potential problems for algorithm compatibility if LARs from opposing forces confront one another. There is also the danger of potential acquisition of LARs by non-State actors, who are less likely to abide by regulatory regimes for control and transparency.
G. Taking human decision-making out of the loop
89. It is an underlying assumption of most legal, moral and other codes that when the decision to take life or to subject people to other grave consequences is at stake, the decision-making power should be exercised by humans. The Hague Convention (IV) requires any combatant “to be commanded by a person”. The Martens Clause, a
65 Ibid , p. 113.
66 A/HR/17/28, p. 17.
67 Asaro (see note 32 above), p. 13.
A/HRC/23/47
17
longstanding and binding rule of IHL, specifically demands the application of “the principle of humanity” in armed conflict.68 Taking humans out of the loop also risks taking humanity out of the loop.
90. According to philosopher Peter Asaro, an implicit requirement can thus be found in IHL for a human decision to use lethal force, which cannot be delegated to an automated process. Non-human decision-making regarding the use of lethal force is, by this argument, inherently arbitrary, and all resulting deaths are arbitrary deprivations of life.69
91. The contemplation of LARs is inextricably linked to the role of technology in the world today. While machines help to make many decisions in modern life, they are mostly so used only where mechanical observation is needed (e.g. as a line umpire in sporting events) and not in situations requiring value judgements with far-reaching consequences (e.g. in the process of adjudication during court cases). As a more general manifestation of the importance of person-to-person contact when important decisions are taken, legal systems around the world shy away from trials in absentia. Of course, robots already affect our lives extensively, including through their impact on life and death issues. Robotic surgery is for example a growing industry and robots are increasingly used in rescue missions after disasters.70 Yet in none of these cases do robots make the decision to kill and in this way LARs represent an entirely new prospect.
92. Even if it is assumed that LARs – especially when they work alongside human beings – could comply with the requirements of IHL, and it can be proven that on average and in the aggregate they will save lives, the question has to be asked whether it is not inherently wrong to let autonomous machines decide who and when to kill. The IHL concerns raised in the above paragraphs relate primarily to the protection of civilians. The question here is whether the deployment of LARs against anyone, including enemy fighters, is in principle acceptable, because it entails non-human entities making the determination to use lethal force.
93. This is an overriding consideration: if the answer is negative, no other consideration can justify the deployment of LARs, no matter the level of technical competence at which they operate. While the argument was made earlier that the deployment of LARs could lead to a vacuum of legal responsibility, the point here is that they could likewise imply a vacuum of moral responsibility.
94. This approach stems from the belief that a human being somewhere has to take the decision to initiate lethal force and as a result internalize (or assume responsibility for) the cost of each life lost in hostilities, as part of a deliberative process of human interaction. This applies even in armed conflict. Delegating this process dehumanizes armed conflict even further and precludes a moment of deliberation in those cases where it may be feasible. Machines lack morality and mortality, and should as a result not have life and death powers over humans. This is among the reasons landmines were banned.71
95. The use of emotive terms such as “killer robots” may well be criticized. However, the strength of the intuitive reactions that the use of LARs is likely to elicit cannot be ignored. Deploying LARs has been depicted as treating people like “vermin”, who are
68 Geneva Convention Protocol I, art. 1(2). See also the preambles to the 1899 and 1907 Hague Conventions. Hague Convention with Respect to the Laws and Customs of War on Land and its Annex: Regulation Concerning the Laws and Customs of War on Land (Hague Convention II)
69 Asaro (see note 43 above), p. 13.
70 See http://www.springer.com/medicine/surgery/journal/11701
71 Asaro (see note 43 above), p. 14.
A/HRC/23/47
18
“exterminated.”72 These descriptions conjure up the image of LARs as some kind of mechanized pesticide.
96. The experience of the two World Wars of the last century may provide insight into the rationale of requiring humans to internalize the costs of armed conflict, and thereby hold themselves and their societies accountable for these costs. After these wars, during which the devastation that could be caused by modern technology became apparent, those who had personally taken the central military decisions resolved, “in order to save succeeding generations from the scourge of war”, to establish the United Nations to pursue world peace and to found it on the principles of human rights. While armed conflict is by no means a thing of the past today, nearly 70 years have passed without a global war. The commitment to achieve such an objective can be understood as a consequence of the long-term and indeed inter-generational effects of insisting on human responsibility for killing decisions.
97. This historical recollection highlights the danger of measuring the performance of LARs against minimum standards set for humans during armed conflict. Human soldiers do bring a capacity for depravity to armed conflict, but they also hold the potential to adhere to higher values and in some cases to show some measure of grace and compassion. If humans are replaced on the battlefield by entities calibrated not to go below what is expected of humans, but which lack the capacity to rise above those minimum standards, we may risk giving up on hope for a better world. The ability to eliminate perceived “troublemakers” anywhere in the world at the press of a button could risk focusing attention only on the symptoms of unwanted situations. It would distract from, or even preclude, engagement with the causes instead, through longer term, often non-military efforts which, although more painstaking, might ultimately be more enduring. LARs could thus create a false sense of security for their users.
H. Other concerns
98. The possible deployment of LARs raises additional concerns that include but are not limited to the following:
• LARs are vulnerable to appropriation, as well as hacking and spoofing.73 States no longer hold a monopoly on the use of force. LARs could be intercepted and used by non-State actors, such as criminal cartels or private individuals, against the State or other non-State actors, including civilians.74
• Malfunctions could occur. Autonomous systems can be “brittle”.75 Unlikely errors can still be catastrophic.
• Future developments in the area of technology cannot be foreseen. Allowing LARs could open an even larger Pandora‟s box.
• The regulation of the use of UCAVs is currently in a state of contestation, as is the legal regime pertaining to targeted killing in general, and the emergence of LARs is likely to make this situation even more uncertain.
• The prospect of being killed by robots could lead to high levels of anxiety among at least the civilian population.
72 Robert Sparrow “Robotic Weapons and the Future of War” in Jessica Wolfendale and Paolo Tripodi (eds.) New Wars and New Soldiers: Military Ethics in the Contemporary World (2011), p. 11.
73 Jutta Weber “Robotic warfare, human rights and the rhetorics of ethical machines”, pp. 8 and 10, available from http://www.gender.uu.se/digitalAssets/44/44133_Weber_Robotic_Warfare.pdf
74 Singer (see note 3 above), p. 261-263.
75 Kastan (see note 55 above), p. 8.
A/HRC/23/47
19
99. The implications for military culture are unknown, and LARs may thus undermine the systems of State and international security.
I. LARs and restrictive regimes on weapons
100. The treaty restrictions76 placed on certain weapons stem from the IHL norm that the means and methods of warfare are not unlimited, and as such there must be restrictions on the rules that determine what weapons are permissible.77 The Martens Clause prohibits weapons that run counter to the “dictates of public conscience.” The obligation not to use weapons that have indiscriminate effects and thus cause unnecessary harm to civilians underlies the prohibition of certain weapons,78 and some weapons have been banned because they “cause superfluous injury or unnecessary suffering”79 to soldiers as well as civilians.80 The use of still others is restricted for similar reasons.81
101. In considering whether restrictions as opposed to an outright ban on LARs would be more appropriate, it should be kept in mind that it may be more difficult to restrict LARs as opposed to other weapons because they are combinations of multiple and often multipurpose technologies. Experts have made strong arguments that a regulatory approach that focuses on technology – namely, the weapons themselves – may be misplaced in the case of LARs and that the focus should rather be on intent or use.82
102. Disarmament law and its associated treaties, however, provide extensive examples of the types of arms control instruments that establish bans or restrictions on use and other activities. These instruments can be broadly characterized as some combination of type of restriction and type of activity restricted. The types of restrictions include a ban or other limitations short of a ban.
103. The type of activity that is typically restricted includes: (i) acquisition, retention or stockpiling, (ii) research (basic or applied) and development, (iii) testing, (iv) deployment, (v) transfer or proliferation, and (vi) use.83
104. Another positive development in the context of disarmament is the inclusion of victim assistance in weapons treaties.84 This concern for victims coincides with other efforts to address the harm weapons and warfare cause to civilians, including the practice of casualty counting85 and the good faith provision of amends – implemented for example by
76 Through the Hague Convention of 1907 and the 1977 Additional Protocols to the Geneva Conventions.
77 See http://www.icrc.org/eng/war-and-law/conduct-hostilities/methods-means-warfare/index.jsp
78 Mine Ban Treaty (1997); and Convention on Cluster Munitions (2008).
79 Protocol I additional to the Geneva Conventions, 1977. art. 35 (2); ICRC, Customary Humanitarian Law, Rule 70.
80 Protocol for the Prohibition of the Use of Asphyxiating, Poisonous or Other Gases, and of Bacteriological Methods of Warfare. Geneva, 17 June 1925.
81 Convention on Certain Conventional Weapons, Protocol III on incendiary weapons.
82 Marchant (see note 26 above), p. 287, Asaro see note 43 above), p. 10.
83 Marchant (see note 26 above), p. 300. See also Bonnie Docherty “The Time is Now: A Historical Argument for a Cluster Munitions Convention” 20 Harvard Human Rights Law Journal (2007), p. 53 for an overview.
84 Mine Ban Treaty (1997), art. 6, and Convention on Certain Conventional Weapons, Protocol V on Explosive Remnants of War (2003), art. 8. The Convention on Cluster Munitions (2008), art. 5 was groundbreaking in placing responsibility on the affected State.
85 S/2012/376, para. 28 (commending inter alia the commitment by the African Union Mission in Somalia).
A/HRC/23/47
20
some International Security Assistance Force States – in the case of civilian deaths in the absence of recognized IHL violations.86 These practices serve to reaffirm the value of life.
105. There are also meaningful soft law instruments that may regulate the emergence of LARs. Examples of relevant soft law instruments in the field of disarmament include codes of conduct, trans-governmental dialogue, information sharing and confidence-building measures and framework conventions.87 In addition, non-governmental organization (NGO) activity and public opinion can serve to induce restrictions on weapons.
106. Article 36 of the First Protocol Additional to the Geneva Conventions is especially relevant, providing that, “in the study, development, acquisition or adoption of a new weapon, means or methods of warfare, a High Contracting Party is under an obligation to determine whether its employment would, in some or all circumstances, be prohibited by this Protocol or by any other rule of international law applicable to the High Contracting Party.”
107. This process is one of internal introspection, not external inspection, and is based on the good faith of the parties.88 The United States, although not a State party, established formal weapons mechanisms review as early as 1947. While States cannot be obliged to disclose the outcomes of their reviews, one way of ensuring greater control over the emergence of new weapons such as LARs will be to encourage them to be more open about the procedure that they follow in Article 36 reviews generally.
108. In 2012 in a Department of Defense Directive, the United States embarked on an important process of self-regulation regarding LARs, recognizing the need for domestic control of their production and deployment, and imposing a form of moratorium.89 The Directive provides that autonomous weapons “shall be designed to allow commanders and operators to exercise appropriate levels of human judgement over the use of force”.90 Specific levels of official approval for the development and fielding of different forms of robots are identified.91 In particular, the Directive bans the development and fielding of LARs unless certain procedures are followed.92 This important initiative by a major potential LARs producer should be commended and may open up opportunities for mobilizing international support for national moratoria.
IV. Conclusions
109. There is clearly a strong case for approaching the possible introduction of LARs with great caution. If used, they could have far-reaching effects on societal values, including fundamentally on the protection and the value of life and on international stability and security. While it is not clear at present how LARs could be capable of satisfying IHL and IHRL requirements in many respects, it is foreseeable that they could comply under certain circumstances, especially if used alongside human soldiers. Even so, there is widespread concern that allowing LARs to kill people may denigrate the value of life itself. Tireless war machines, ready for deployment at the push of a button, pose the danger of permanent (if low-level) armed conflict, obviating the opportunity for post-war reconstruction. The onus is on those who wish to deploy LARs to demonstrate that specific uses should in particular
86 Ibid., para. 29 (the Secretary General “welcomed the practice of making amends”).
87 Marchant, see note 26, pp. 306-314.
88 Discussed in International Review of the Red Cross vol. 88, December 2006.
89 US DoD Directive (see note 14 above).
90 Ibid, para 4.a.
91 Ibid, paras 4.c and d.
92 Ibid, Enclosure 3.
A/HRC/23/47
21
circumstances be permitted. Given the far-reaching implications for protection of life, considerable proof will be required.
110. If left too long to its own devices, the matter will, quite literally, be taken out of human hands. Moreover, coming on the heels of the problematic use and contested justifications for drones and targeted killing, LARs may seriously undermine the ability of the international legal system to preserve a minimum world order.
111. Some actions need to be taken immediately, while others can follow afterwards. If the experience with drones is an indication, it will be important to ensure that transparency, accountability and the rule of law are placed on the agenda from the start. Moratoria are needed to prevent steps from being taken that may be difficult to reverse later, while an inclusive process to decide how to approach this issue should occur simultaneously at the domestic, intra-State, and international levels.
112. To initiate this process an international body should be established to monitor the situation and articulate the options for the longer term. The ongoing engagement of this body, or a successor, with the issues presented by LARs will be essential, in view of the constant evolution of technology and to ensure protection of the right to life – to prevent both individual cases of arbitrary deprivation of life as well as the devaluing of life on a wider scale.
V. Recommendations
A. To the United Nations
113. The Human Rights Council should call on all States to declare and implement national moratoria on at least the testing, production, assembly, transfer, acquisition, deployment and use of LARs until such time as an internationally agreed upon framework on the future of LARs has been established;
114. Invite the High Commissioner for Human Rights to convene, as a matter of priority, a High Level Panel on LARs consisting of experts from different fields such as law, robotics, computer science, military operations, diplomacy, conflict management, ethics and philosophy. The Panel should publish its report within a year, and its mandate should include the following:
(a) Take stock of technical advances of relevance to LARs;
(b) Evaluate the legal, ethical and policy issues related to LARs;
(c) Propose a framework to enable the international community to address effectively the legal and policy issues arising in relation to LARs, and make concrete substantive and procedural recommendations in that regard; in its work the Panel should endeavour to facilitate a broad-based international dialogue;
(d) Assessment of the adequacy or shortcomings of existing international and domestic legal frameworks governing LARs;
(e) Suggestions of appropriate ways to follow up on its work.
115. All relevant United Nations agencies and bodies should, where appropriate in their interaction with parties that are active in the field of robotic weapons:
(a) Emphasize the need for full transparency regarding all aspects of the development of robotic weapon systems;
A/HRC/23/47
22
(b) Seek more international transparency from States regarding their internal weapons review processes, including those under article 36 of Additional Protocol I to the Geneva Conventions.
B. To regional and other inter-governmental organizations
116. Support the proposals outlined in the recommendations to the United Nations and States, in particular the call for moratoria as an immediate step.
117. Where appropriate take similar or parallel initiatives to those of the United Nations.
C. To States
118. Place a national moratorium on LARs as described in paragraph 114.
119. Declare – unilaterally and through multilateral fora – a commitment to abide by IHL and IHRL in all activities surrounding robotic weapons and put in place and implement rigorous processes to ensure compliance at all stages of development.
120. Commit to being as transparent as possible about internal weapons review processes, including metrics used to test robotic systems. States should at a minimum provide the international community with transparency regarding the processes they follow (if not the substantive outcomes) and commit to making the reviews as robust as possible.
121. Participate in international debate and trans-governmental dialogue on the issue of LARs and be prepared to exchange best practices with other States, and collaborate with the High Level Panel on LARs.
D. To developers of robotic systems
122. Establish a code or codes of conduct, ethics and/or practice defining responsible behaviour with respect to LARs in accordance with IHL and IHRL, or strengthen existing ones.
E. To NGOs, civil society and human rights groups and the ICRC
123. Consider the implications of LARs for human rights and for those in situations of armed conflict, and raise awareness about the issue.
124. Assist and engage with States wherever possible in aligning their relevant procedures and activities with IHL and IHRL.
125. Urge States to be as transparent as possible in respect of their weapons review processes.
126. Support the work of the High Level Panel on LARs.

Goodbye Democracy – article in Heddwch first published in 2008

I wrote the following article for Heddwch, the newsletter for CND Cymru, in 2008.  I republish it here for those who may have missed it first time around.  Parc Aberporth is still largely unoccupied and Selex have pulled out.  It is an economic development fiasco but the issues raised vis a vis drone technology are more relevant than ever.

Goodbye Democracy

UAVs over West Wales

‘Skynet’ is a fictional, computer-based military defence system that acts as the
primary antagonist in the ‘Terminator’ series of films and computer games. This
is a fictional example of an artificial intelligence that becomes sentient, and turns
against its creators. Of course this is Science Fiction and not happening in the
real world. Or is it?

An ASTRAEA (Autonomous Systems Technology Related Airborne Evaluation and Assessment) Centre is being established at Parc Aberporth on the Ceredigion coast, with the backing of £3 million of public money on top of the cost of building the Parc itself. ASTRAEA is a British programme with Welsh Assembly and Westminster Governments support and funding, which involves a number of arms companies including the European Aeronautic Defence and Space Company (EADS), QinetiQ, Flight Refuelling Ltd (FRL), Thales and BAE Systems.

Remote killing

At Aberporth, Unmanned Aerial Vehicles (UAVs or drones) are being developed for a variety of purposes, some benign and some most certainly not.  QinetiQ is currently carrying out experiments marrying up ‘artificial intelligence’ with Unmanned Aerial Vehicles (UAVs). While these could have peaceful applications, it quickly becomes worrying when the same technology is applied to Unmanned Combat Aerial Vehicles (UCAVs). The future of warfare is computer controlled and where the human element is removed from the theatre of operations. Civilian populations could be kept in check by a variety of robotic drones.  Some drones are designed to be spy planes for collecting information and intelligence; others carry weapons for combat and bomb delivery. All this is talked up with glee by various politicians and civil servants as a perfect way to solve the problems of economic regeneration in West Wales.

Think again

There are dark things going on at Parc Aberporth. Why is the Welsh Assembly Government so keen to get into bed with these major armaments developers? QinetiQ’s website makes no bones about the fact that the company is linked in with the UCAV programme for the Ministry of Defence and will be connecting in with BAE, joint partners in the ASTRAEA project with offices in the Parc Aberporth complex.
Those in favour of the civil aviation authority airport planned for Aberporth might think twice about arms dealers and international weaponry experts flying in and out of a West Wales Airport. It is time that New Labour ’s Andrew Davies AM, Welsh Assembly Government Minister for Finance and Public Service Delivery, began to understand that Wales is a fledgling democracy and that the people of Wales do not want to be blinded with spin in the usual smoke and mirrors approach of the Westminster style of government.

Despite this huge ‘investment’ it was reported last May that the Parc Aberporth remained largely unoccupied more than 18 months after it was first opened. The Assembly Government had originally announced over 230 jobs would be created by 2008 with a further capacity for up to 1,000 employees. Selex (Sensors and Airborne Systems),  Europe’s second largest defence electronics business occupies one unit on the site. The ‘expertise’ of EADS, QinetiQ, Flight Refuelling Ltd, Thales and BAE Systems was to have been combined with that of university departments at Aberystwyth and Cardiff ’ to create even more jobs. The Parc is now frequented by practicing learner drivers and the grounds are unkempt. As we know, the ideal situation for profit in the military industrial complex is ‘endless war ’.

People in Wales demand better and want to know that industry in Wales is being used for purposes other than being part of the US led world domination inspired military transformation strategy which George W Bush, Tony Blair, Rhodri Morgan, and now, Gordon Brown have bought in to.

Harry Rogers

The Chilly Dogz – “Ray Bradbury Said”

harri seeing red

This is the latest Chilly Dogz Tuesday session.  Every week we meet at my house and write a new song together.  I wrote this after seeing a documentary about deceased Sci Fi author Ray Bradbury in which he said that he never bothered to carry out research for his stories as all the information he needed was in his head.  His fiction is fabulous and has been a favourite of mine for more than 50 years.

LYRIC:-

RAY BRADBURY SAID

IT’S ALL IN MY HEAD

RAY BRADBURY SAID

IT’S ALL IN HIS HEAD

 

HE TALKED IN TECHNICOLOR

NIGHT AFTER NIGHT AFTER NIGHT

HE TRANSFORMED LANGUAGE

INTO SOMETHING MEGA BRIGHT

HE COULD SEE THE AIR WE BREATHE

AND HE TOLD US ALL ABOUT IT

SHIMMERING WORLDS INSIDE HIS HEAD

HE MADE SURE WE KNEW ABOUT IT

 

RAY BRADBURY SAID

IT’S ALL IN MY HEAD

RAY BRADBURY SAID

IT’S ALL…

View original post 61 more words

One Way Ticket To Mars

ONE WAY TICKET TO MARS

I had a ginger cat that sometimes killed a mouse

I had a shiny car parked up outside my house

I had a steady job that I really thought would last

Now everything has gone, relegated to the past

My dreams and aspirations have all come tumbling down

Wife and kids live with her mother in another part of town

The only thing that’s left now I’ve landed on my arse

Is to try and get myself a one way ticket to Mars

A place where I’ll be blinded by the brightness of the stars

Where there will be no driving round in comfortable cars

No more late night drinking in my favourite bars

No more gigs down Deptford with those sweet, sweet guitars

It’s time I got myself on that one way ticket to Mars

There are no rotten human beings there to treat me so unkind

All those problems down on Planet Earth?  I’ll leave them all behind

I’ll climb into that rocket ship and become a pioneer

Things can’t be any worse up there than all this shit down here

I’ve got no cash inside my pocket

So strap me up into that rocket

I’ll take that one way ticket to Mars

I wanna one way ticket to Mars

I’m on a one way ticket to Mars

A One…….. Way……. Ticket

See ya……wouldn’t wanna be ya!

http://www.bbc.co.uk/news/science-environment-22146456#

NaPoWriMo April 2013 # 23 – Poker Alice

POKER ALICE

COLORADO SAW A BIG GAME

POKER ALICE V SOAPY SMITH

SHE WON SIX THOUSAND DOLLARS

WITH A SEVEN AND A DEUCE

SOAPY WASN’T HAPPY

AS THE FULL HOUSE WAS REVEALED

ALICE LEFT THE WHITE ELEPHANT

TURNED HER WHOREHOUSE LOOSE

 

ALL THE MINERS AND THE GAMBLERS

THOUGHT IT WAS A SPECIAL TIME

THEY RAN OUT OF ALL THE SALOONS

SHOUTING HIP HIP HOORAY

SHE CHAWED MORE THAN THREE TIMES

ON HER BIG CUBAN CIGAR

THIS WAS A LOT MORE MONEY

THAN SHE’D EVER WON BY FAR

 

POKER ALICE

SHE WAS A WILD GIRL

POKER ALICE

A WILD WILD GIRL

 

ALICE OFTEN SAID AS

SHE RAKED HER WINNINGS IN

“PRAISE THE LORD AND PLACE YOUR BETS.

I’LL TAKE YOUR MONEY WITH NO REGRETS.”

YOU NEVER FOUND ALICE KNITTING

SHE NEVER SAT MAKING LACE

SHE’D RATHER SPEND HER TIME

PUTTING ON HER POKER FACE

 

POKER ALICE

SHE WAS A WILD GIRL

POKER ALICE

A WILD WILD GIRL

NaPoWriMo April 2013 # 22 – Kismet Trouve

KISMET TROUVE

Dry leaves blow across the new mown grass

Swirling around and around and around

Just the same as those thoughts in my head

Swirling around and around and around

 

I just cannot tell what is happening to me

Something deep inside me seems to be stirring

My heart is aching, and my brain is burning

Constantly consumed by whizzing and whirring

 

Never felt this before

It’s Kismet trouvé

Never felt this before

It’s Kismet trouvé

 

The river roars into the whirlpool over the falls

Swirling around and around and around

A myriad of feelings fill my body and my mind

Swirling around and around and around

 

Giant icicles are melting under Henllan Bridge

The sun is burning off the early morning mist

A high pitched Buzzard calls way above the ridge

Things aint the same since that moment we kissed

 

Never felt this before

It’s Kismet trouvé

Never felt this before

It’s Kismet trouvé

 

I’ve found my fate

Before it’s too late

NaPoWriMo 2013 #10 – Baby It’s Cold Inside

NaPoWriMo 2013 poem number 10 written for the daughter of a dear friend.

BABY IT’S COLD INSIDE


SHE USED TO BE ABLE

TO SPEND HALF HER DAY

SAT AT HER TABLE

OUTSIDE THE TIME CAFE

WITH A CAPPUCINO

AND A ROLL UP CIGARETTE

WATCHING THE WORLD GO

HOW MUCH BETTER CAN IT GET?

 

SHE’S TRAPPED NOW

BY THE WHITE COATS

SHE’S TRAPPED NOW

BY THE WHITE COATS

 

SHE WANTS TO GO OUT

LIKE SHE ALWAYS HAD

SHE WANTS TO GO OUT

THEY SAY SHE IS MAD

KEEP HER LOCKED AWAY

IN THAT WAY THEY CAN

LOCK HER WELL AWAY

IN COLD COMFORT LAND

 

SHE’S TRAPPED NOW

BY THE WHITE COATS

SHE’S TRAPPED NOW

BY THE WHITE COATS

 

THEY TOOK HER FREEDOM

WRAPPED IT UP IN RULES

RULES OVER FREEDOM

FROM A BUNCH OF FOOLS

SAY IT’S FOR THE BEST

FOR HER PROTECTION

IT’S THE VERY BEST

FORWARD PROJECTION

 

SHE’S TRAPPED NOW

BY THE WHITE COATS

SHE’S TRAPPED NOW

BY THE WHITE COATS

 

BABY IT’S COLD INSIDE

BABY IT’S COLD INSIDE

 

Copyright: 05-03-2013

National Poetry Writing Month

I must be crazy but I am taking part in NaPoWriMo, or National Poetry Writing Month, an annual project in which participating poets attempt to write a poem a day for the month of April. My first poem will appear here on 1st April and I hope I can keep the pencil moving until the end of the month.  This is a challenge and I have no idea what I will be writing about but hey, that’s the beauty of the thing. Check out the NaPoWriMo site at http://www.napowrimo.net/ or see it on Facebook at https://www.facebook.com/pages/National-Poetry-Writing-Month-2013/415972748492192 .