WHAT IS INTRO TO DATA SCIENCE: ARTIFICIAL INTELLIGENCE—AT THE INTERSECTION OF CS AND DATA SCIENCE IN PYTHON PROGRAMMING

INTRO TO DATA SCIENCE: ARTIFICIAL INTELLIGENCE—AT THE INTERSECTION OF CS AND DATA SCIENCE

    When a baby first opens its eyes, does it “see” its parent’s faces? Does it understand any notion of what a face is—or even what a simple shape is? Babies must “learn” the world around them. That’s what artificial intelligence (AI) is doing today. It’s looking at massive amounts of data and learning from it. AI is being used to play games, implement a wide range of computer­vision applications, enable self­driving cars, enable robots to learn to perform new tasks, diagnose medical conditions, translate speech to other languages in near real time, create chatbots that can respond to arbitrary questions using massive databases of knowledge, and much more. Who’d have guessed just a few years ago that artificially intelligent self­driving cars would be allowed on our roads—or even become common? Yet, this is now a highly competitive area. The ultimate goal of all this learning is artificial general intelligence—an AI that can perform intelligence tasks as well as humans. This is a scary thought to many people.

      Artificial-Intelligence Milestones

        Several artificial­-intelligence milestones, in particular, captured people’s attention and imagination, made the general public start thinking that AI is real and made businesses think about commercializing AI:
        • In a 1997 match between IBM’s DeepBlue computer system and chess Grandmaster Gary Kasparov, DeepBlue became the first computer to beat a reigning world chess champion under tournament conditions. IBM loaded DeepBlue with hundreds of thousands of grandmaster chess games. DeepBlue was capable of using brute force to evaluate up to 200 million moves per second! This is big data at work. IBM received the Carnegie Mellon University Fredkin Prize, which in 1980 offered $100,000 to the creators of the first computer to beat a world chess champion.
          1. https://en.wikipedia.org/wiki/Deep_Blue_versus_Garry_Kasparov.
          2. https://en.wikipedia.org/wiki/Deep_Blue_(chess_computer).
          • In 2011, IBM’s Watson beat the two best human Jeopardy! players in a $1 million match. Watson simultaneously used hundreds of language­analysis techniques to locate correct answers in 200 million pages of content (including all of Wikipedia) requiring four terabytes of storage. Watson was trained with machine learning and reinforcement­-learning techniques. Chapter 13 discusses IBM Watson and Chapter 14 discusses machine­learning.
          • Go—a board game created in China thousands of years ago —is widely considered to be one of the most complex games ever invented with 10 possible board configurations. To give you a sense of how large a number that is, it’s believed that there are (only) between 10 and 10 atoms in the known universe! In 2015, AlphaGo—created by Google’s DeepMind group—used deep learning with two neural networks to beat the European Go champion Fan Hui. Go is considered to be a far more complex game than chess. Chapter 15 discusses neural networks and deep learning.
          • More recently, Google generalized its AlphaGo AI to create AlphaZero—a game­playing AI that teaches itself to play other games. In December 2017, AlphaZero learned the rules of and taught itself to play chess in less than four hours using reinforcement learning. It then beat the world champion chess program, Stockfish 8, in a 100­game match—winning or drawing every game. After training itself in Go for just eight hours, AlphaZero was able to play Go vs. its AlphaGo predecessor, winning 60 of 100 games.
          1. https://www.theguardian.com/technology/2017/dec/07/alphazero-google-deepmind-ai-beats-champion-program-teaching-itself-to-play-four-hours.

          A Personal Anecdote

          When one of the authors, Harvey Deitel, was an undergraduate student at MIT in the mid-1960s, he took a graduate­level artificial­intelligence course with Marvin Minsky (to whom this article is dedicated), one of the founders of artificial intelligence (AI). Harvey: Professor Minsky required a major term project. He told us to think about what intelligence is and to make a computer do something intelligent. Our grade in the course would be almost solely dependent on the project. No pressure! I researched the standardized IQ tests that schools administer to help evaluate their students’ intelligence capabilities. Being a mathematician at heart, I decided to tackle the popular IQ­test problem of predicting the next number in a sequence of numbers of arbitrary length and complexity. I used interactive Lisp running on an early Digital Equipment Corporation PDP­1 and was able to get my sequence predictor running on some pretty complex stuf , handling challenges well beyond what I recalled seeing on IQ tests. Lisp’s ability to manipulate arbitrarily long lists recursively was exactly what I needed to meet the project’s requirements. Python of ers recursion and generalized list processing (Chapter 5).
          I tried the sequence predictor on many of my MIT classmates. They would make up numbersequences and type them into my predictor. The PDP­1 would “think” for a while—often a long while—and almost always came up with the right answer. Then I hit a snag. One of my classmates typed in the sequence 14, 23, 34 and 42. My predictor went to work on it, and the PDP­1 chugged away for a long time, failing to predict the next number. I couldn’t get it either. My classmate told me to think about it overnight, and he’d reveal the answer the next day, claiming that it was a simple sequence. My efforts were to no avail. The following day he told me the next number was 57, but I didn’t understand why. So he told me to think about it overnight again, and the following day he said the next number was 125. That didn’t help a bit—I was stumped. He said that the sequence was the numbers of the two­way crosstown streets of Manhattan. I cried, “foul,” but he said it met my criterion of predicting the next number in a numerical sequence. My world view was mathematics—his was broader. Over the years, I’ve tried that sequence on friends, relatives and professional colleagues. A few who spent time in Manhattan got it right. My sequence predictor needed a lot more than just mathematical knowledge to handle problems like this, requiring (a possibly vast) world knowledge.

          Watson and Big Data Open New Possibilities

          When Paul and I started working on this Python article, we were immediately drawn to IBM’s Watson using big data and artificial­intelligence techniques like natural language processing (NLP) and machine learning to beat two of the world’s best human Jeopardy! players. We realized that Watson could probably handle problems like the sequence predictor because it was loaded with the world’s street maps and a whole lot more. That chet our appetite for digging in deep on big data and today’s artificial­intelligence technologies, and helped shape Chapters 11–16 of this article. It’s notable that all of the data­science implementation case studies in Chapters 11–16 either are rooted in artificial intelligence technologies or discuss the big data hardware and software infrastructure that enables computer scientists and data scientists to implement leading­-edge AI-based solutions effectively.

          AI: A Field with Problems But No Solutions

          For many decades, AI has been viewed as a field with problems but no solutions. That’s because once a particular problem is solved people say, “Well, that’s not intelligence, it’s just a computer program that tells the computer exactly what to do.” However, with machine learning (Chapter 14) and deep learning (Chapter 15) we’re not pre­programming­ solutions to specific problems. Instead, we’re letting our computers solve problems by learning from data—and, typically, lots of it. Many of the most interesting and challenging problems are being pursued with deep learning. Google alone has thousands of deep­learning projects underway and that number is growing quickly. As you work through this article, we’ll introduce you to many edge­-of-the-­practice artificial intelligence, big data and cloud technologies.

          *

          إرسال تعليق (0)
          أحدث أقدم