One of the most intriguing books I’ve read this year is Machinehood by S.B. Divya. Set in the year 2095, AI and bots have taken over most of the work that humans once did, leaving humans little to do but supervise the bots or work in the gig economy, doing the very few tasks that AI can’t yet handle. It’s a dystopian future, driven by a capitalist drive to increase revenues and profits without regard to the human costs of replacing humanity with machines.
How do we avoid an AI dystopia as envisioned in Machinehood? One approach is to leave it to government regulation. Although various countries are considering such legislation, and I applaud their efforts, I think it’s unlikely that they will strike the right balance between protecting their citizenry and innovation. I’m also skeptical that they will put together regulations in a timely fashion, keeping up with technological changes. Another approach, and one that I favor, is for organizations to create their own AI policies. If enough organizations do this and adopt the best of what their peers have done, we may get something both responsible and workable.
Here are a few examples of AI policies:
Randall Pine: https://www.randallpine.com/generative-ai-policy-template
Fisher Phillips: https://www.fisherphillips.com/a/web/du6wach1kmRuPCgDcMLJ5Z/ai-policy.pdf
Wired Magazine: https://www.wired.com/about/generative-ai-policy/
In looking at these AI policies, a few trends and thoughts come to mind:
Melissa and I at the Agile Marketing Alliance are working to create our own organizational AI policy. Does your organization have an AI policy? Can you share an external version of it?
Content Authenticity Statement: 100% of this content was generated by me, the human. I did use Grammarly to check spelling and grammar.
Your partner in achieving improved business outcomes faster with Marketing AI consulting.