econintersect.com
  • 토토사이트
    • 카지노사이트
    • 도박사이트
    • 룰렛 사이트
    • 라이브카지노
    • 바카라사이트
    • 안전카지노
  • 경제
  • 파이낸스
  • 정치
  • 투자
No Result
View All Result
  • 토토사이트
    • 카지노사이트
    • 도박사이트
    • 룰렛 사이트
    • 라이브카지노
    • 바카라사이트
    • 안전카지노
  • 경제
  • 파이낸스
  • 정치
  • 투자
No Result
View All Result
econintersect.com
No Result
View All Result
Home Uncategorized

Artificial Intelligence Can Now Emulate Human Behaviors – Soon It Will Be Dangerously Good

admin by admin
9월 6, 2021
in Uncategorized
0
0
SHARES
0
VIEWS

from The Conversation

— this post authored by Ana Santos Rutschman, Saint Louis University

When artificial intelligence systems start getting creative, they can create great things – and scary ones. Take, for instance, an AI program that let web users compose music along with a virtual Johann Sebastian Bach by entering notes into a program that generates Bach-like harmonies to match them.


Please share this article – Go to very top of page, right hand side, for social media buttons.


Run by Google, the app drew great praise for being groundbreaking and fun to play with. It also attracted criticism, and raised concerns about AI’s dangers.

My study of how emerging technologies affect people’s lives has taught me that the problems go beyond the admittedly large concern about whether algorithms can really create music or art in general. Some complaints seemed small, but really weren’t, like observations that Google’s AI was breaking basic rules of music composition.

In fact, efforts to have computers mimic the behavior of actual people can be confusing and potentially harmful.

Impersonation technologies

Google’s program analyzed the notes in 306 of Bach’s musical works, finding relationships between the melody and the notes that provided the harmony. Because Bach followed strict rules of composition, the program was effectively learning those rules, so it could apply them when users provided their own notes.

The Google Doodle team explains the Bach program.

The Bach app itself is new, but the underlying technology is not. Algorithms trained to recognize patterns and make probabilistic decisions have existed for a long time. Some of these algorithms are so complex that people don’t always understand how they make decisions or produce a particular outcome.

AI systems are not perfect – many of them rely on data that aren’t representative of the whole population, or that are influenced by human biases. It’s not entirely clear who might be legally responsible when an AI system makes an error or causes a problem.

Now, though, artificial intelligence technologies are getting advanced enough to be able to approximate individuals’ writing or speaking style, and even facial expressions. This isn’t always bad: A fairly simple AI gave Stephen Hawking the ability to communicate more efficiently with others by predicting the words he would use the most.

More complex programs that mimic human voices assist people with disabilities – but can also be used to deceive listeners. For example, the makers of Lyrebird, a voice-mimicking program, have released a simulated conversation between Barack Obama, Donald Trump and Hillary Clinton. It may sound real, but that exchange never happened.

From good to bad

In February 2019, nonprofit company OpenAI created a program that generates text that is virtually indistinguishable from text written by people. It can “write” a speech in the style of John F. Kennedy, J.R.R. Tolkien in “The Lord of the Rings” or a student writing a school assignment about the U.S. Civil War.

The text generated by OpenAI’s software is so believable that the company has chosen not to release the program itself.

Similar technologies can simulate photos and videos. In early 2018, for instance, actor and filmmaker Jordan Peele created a video that appeared to show former U.S. President Barack Obama saying things Obama never actually said to warn the public about the dangers posed by these technologies.

Be careful what videos you believe.

In early 2019, a fake nude photo of U.S. Rep. Alexandria Ocasio-Cortez circulated online. Fabricated videos, often called “deepfakes,” are expected to be increasingly used in election campaigns.

Members of Congress have started to look into this issue ahead of the 2020 election. The U.S. Defense Department is teaching the public how to spot doctored videos and audio. News organizations like Reuters are beginning to train journalists to spot deepfakes.

But, in my view, an even bigger concern remains: Users might not be able to learn fast enough to distinguish fake content as AI technology becomes more sophisticated. For instance, as the public is beginning to become aware of deepfakes, AI is already being used for even more advanced deceptions. There are now programs that can generate fake faces and fake digital fingerprints, effectively creating the information needed to fabricate an entire person – at least in corporate or government records.

Machines keep learning

At the moment, there are enough potential errors in these technologies to give people a chance of detecting digital fabrications. Google’s Bach composer made some mistakes an expert could detect. For example, when I tried it, the program allowed me to enter parallel fifths, a music interval that Bach studiously avoided. The app also broke musical rules of counterpoint by harmonizing melodies in the wrong key. Similarly, OpenAI’s text-generating program occasionally wrote phrases like “fires happening under water” that made no sense in their contexts.

As developers work on their creations, these mistakes will become rarer. Effectively, AI technologies will evolve and learn. The improved performance has the potential to bring many social benefits – including better health care, as AI programs help democratize the practice of medicine.

Giving researchers and companies freedom to explore, in order to seek these positive achievements from AI systems, means opening up the risk of developing more advanced ways to create deception and other social problems. Severely limiting AI research could curb that progress. But giving beneficial technologies room to grow comes at no small cost – and the potential for misuse, whether to make inaccurate “Bach-like” music or to deceive millions, is likely to grow in ways people can’t yet anticipate.The Conversation

Ana Santos Rutschman, Assistant Professor of Law, Saint Louis University

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Previous Post

Wine-Food Pairings: Done The Right Way At The Lenox Club

Next Post

What Jobless Claims Mean For Stock Market Returns and Employment

Related Posts

Scammers Steal $300K Using Fake Blur Airdrop Websites
Uncategorized

FBI Warns Investors Of Crypto-Stealing Play-to-Earn Games

by admin
Maersk Almost Completing Russia Exit After The Sale Of Logistics Sites
Uncategorized

Maersk Almost Completing Russia Exit After The Sale Of Logistics Sites

by admin
Why Is ‘Staking’ At The Center Of Crypto’s Latest Regulation Scuffle
Uncategorized

Why Is ‘Staking’ At The Center Of Crypto’s Latest Regulation Scuffle

by admin
Mexico's Pemex Dismantled Resources Worth $342M From Two Top Fields
Uncategorized

Mexico’s Pemex Dismantled Resources Worth $342M From Two Top Fields

by admin
Oil Giant Schlumberger Rebrands Itself As SLB For Low-Carbon Future
Uncategorized

Oil Giant Schlumberger Rebrands Itself As SLB For Low-Carbon Future

by admin
Next Post

Democratic Governors Are Quicker In Responding To The Coronavirus Than Republicans

답글 남기기 응답 취소

이메일 주소는 공개되지 않습니다. 필수 필드는 *로 표시됩니다

Browse by Category

  • Business
  • Econ Intersect News
  • Economics
  • Finance
  • Politics
  • Uncategorized

Browse by Tags

adoption altcoins bank banking banks Binance Bitcoin Bitcoin market blockchain BTC BTC price business China crypto crypto adoption cryptocurrency crypto exchange crypto market crypto regulation decentralized finance DeFi Elon Musk ETH Ethereum Europe Federal Reserve finance FTX inflation investment market analysis Metaverse NFT nonfungible tokens oil market price analysis recession regulation Russia stock market technology Tesla the UK the US Twitter

Categories

  • Business
  • Econ Intersect News
  • Economics
  • Finance
  • Politics
  • Uncategorized

© Copyright 2024 EconIntersect

No Result
View All Result
  • 토토사이트
    • 카지노사이트
    • 도박사이트
    • 룰렛 사이트
    • 라이브카지노
    • 바카라사이트
    • 안전카지노
  • 경제
  • 파이낸스
  • 정치
  • 투자

© Copyright 2024 EconIntersect