Google: Our AI Wont be Used As Weapons

"Don't Be Evil" they said - as they were secretly recording and monitoring us....

 

Now they say their AI won't be weaponized. Do you believe them? As it is - Google already has plenty of shady business practices such as recording personal conversations, tracking people's movement, storing and analysing everything you have ever searched for on the internet, and pretending to be search engine for the last 20 years while quietly amassing a storehouse of personal information, habits and spending patterns of not just Americans but of people all across the globe.

Won't use it as a weapon? Me thinks thou protest too much - and it's a little late by the way. From what I can tell , Google has already been weaponized.

The silicon valley giant announced through it's AI principles that it would no longer " pursue weapons or other technologies whose principal purpose is to cause harm. "

Here is an excerpt from the paper just released hours ago:

Objectives for AI applications

We will assess AI applications in view of the following objectives. We believe that AI should:

 

1) Be socially beneficial.

The expanded reach of new technologies increasingly touch society as a whole. Advances in AI will have transformative impacts in a wide range of fields, including healthcare, security, energy, transportation, manufacturing, and entertainment. As we consider potential development and uses of AI technologies, we will take into account a broad range of social and economic factors, and will proceed where we believe that the overall likely benefits substantially exceed the foreseeable risks and downsides.

AI also enhances our ability to understand the meaning of content at scale. We will strive to make high-quality and accurate information readily available using AI, while continuing to respect cultural, social, and legal norms in the countries where we operate. And we will continue to thoughtfully evaluate when to make our technologies available on a non-commercial basis.

2)Avoid creating or reinforcing unfair bias.

AI algorithms and datasets can reflect, reinforce, or reduce unfair biases. We recognize that distinguishing fair from unfair biases is not always simple, and differs across cultures and societies. We will seek to avoid unjust impacts on people, particularly those related to sensitive characteristics such as race, ethnicity, gender, nationality, income, sexual orientation, ability, and political or religious belief.

3) Be built and tested for safety.

We will continue to develop and apply strong safety and security practices to avoid unintended results that create risks of harm. We will design our AI systems to be appropriately cautious, and seek to develop them in accordance with best practices in AI safety research. In appropriate cases, we will test AI technologies in constrained environments and monitor their operation after deployment.

4) Be accountable to people.

We will design AI systems that provide appropriate opportunities for feedback, relevant explanations, and appeal. Our AI technologies will be subject to appropriate human direction and control.

5) Incorporate privacy design principles.

We will incorporate our privacy principles in the development and use of our AI technologies. We will give opportunity for notice and consent, encourage architectures with privacy safeguards, and provide appropriate transparency and control over the use of data.

6) Uphold high standards of scientific excellence.

Technological innovation is rooted in the scientific method and a commitment to open inquiry, intellectual rigor, integrity, and collaboration. AI tools have the potential to unlock new realms of scientific research and knowledge in critical domains like biology, chemistry, medicine, and environmental sciences. We aspire to high standards of scientific excellence as we work to progress AI development.

We will work with a range of stakeholders to promote thoughtful leadership in this area, drawing on scientifically rigorous and multidisciplinary approaches. And we will responsibly share AI knowledge by publishing educational materials, best practices, and research that enable more people to develop useful AI applications.

7) Be made available for uses that accord with these principles.

Many technologies have multiple uses. We will work to limit potentially harmful or abusive applications. As we develop and deploy AI technologies, we will evaluate likely uses in light of the following factors:

  • Primary purpose and use: the primary purpose and likely use of a technology and application, including how closely the solution is related to or adaptable to a harmful use

  • Nature and uniqueness: whether we are making available technology that is unique or more generally available

  • Scale: whether the use of this technology will have significant impact

  • Nature of Google’s involvement: whether we are providing general-purpose tools, integrating tools for customers, or developing custom solutions

Along with a list of Do's - they have also put out list Do Not's - things that they will NOT be pursuing

 

from the release:

AI applications we will not pursue

In addition to the above objectives, we will not design or deploy AI in the following application areas:

  1. Technologies that cause or are likely to cause overall harm. Where there is a material risk of harm, we will proceed only where we believe that the benefits substantially outweigh the risks, and will incorporate appropriate safety constraints.

  2. Weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people.

  3. Technologies that gather or use information for surveillance violating internationally accepted norms.

  4. Technologies whose purpose contravenes widely accepted principles of international law and human rights.

We want to be clear that while we are not developing AI for use in weapons, we will continue our work with governments and the military in many other areas. These include cybersecurity, training, military recruitment, veterans’ healthcare, and search and rescue. These collaborations are important and we’ll actively look for more ways to augment the critical work of these organizations and keep service members and civilians safe.

 

So they still plan to work very closely with the government for cybersecurity , training, and military recruitment. What do you think it would look like for AI to be involved with military recruitment? Is it going to scour the web looking for people who seem like they might be patriotic and target them as potential soldiers? Would it use AI to convince people who might be on the fence about joining? It's very interesting to think about what role AI could play in military recruitment.

The whole thing sounds a little too altruistic to be true - and there is no reason to think that they wouldn't change their mind in the future. By the way - they have removed the "Don't Be Evil" clause from their mantra.

What do you think - do you believe that the most powerful company on the planet will wield it's power for good instead of evil?

Article Source:

https://www.blog.google/topics/ai/ai-principles/

 
 


 
 

 

Please be sure to check out my others articles on real estate, investing , and cryptocurrency ( and some other random dog stuff)

I always upvote comments to my blog posts

 

Solar Panels to Become Mandatory in California

Are Steemit Resteem Bots a Let Down in Most Cases?

Never Give Up - Today is Your Day

Lake Tahoe: Finding Peace in the Serenity of Mountain Living

Experts say to ask your baby for consent before changing diapers now

Taming the Wild West: Building our First Fire Pit

From Trash to Trellis: Beating the Big Box Stores with Innovation

Desolate Wasteland Transformed into Vibrant Colorful Garden

H2
H3
H4
3 columns
2 columns
1 column
Join the conversation now
Logo
Center