Categories
Uncategorized

AIDL Weekly #13 – Special Issue on GTC 2017

Editorial

The Moat of Nvidia – Thoughts From Your Humble Curators

There are many tech conferences each year. But none impressed us as much as GTC 2017. We curated 4 pieces about the conference, but in this Editorial, we’d to explain the incredible moat of Nvidia. And, we think this moat is getting stronger.

First, by “moat”, we mean competitive advantage. So what’s Nvidia’s moat? Some of you might quickly point out its hardware platforms such as its GTX, Quadro and Teslas (or Pascal or Volta) series of GPU cards, and software platform, CUDA. Beyond the obvious IP and chip design moat, there is also powerful software lock-in. Indeed, as developers, we compile code with CUDA daily. CUDA is an easy to learn extension of C and is quick to produce results. The surrounding rich software support makes it easy to get up and running, and has high switching costs, once enough efforts has been spent on top of it.

But increasingly, Nvidia is branching out into new areas of computing, creating new moats. It just tripled its data center business in a yoy basis. It has to do with the fact that they own both the hardware/software platform. And deep learning is not going anywhere soon.

Now, this moat is further strengthening in GTC 2017. Why? First, it announced that it is going to train 100k developers just this year, creating more potential customers steeped in their wares. This is a smart move – behaviors are hard to change. Secondly, they announced a new cloud platform initiative (curated under “Nvidia GPU Cloud”), which makes it easier for newcomers to start building on Nvidia’s platform. Now, it remains to be seen what the competitive dynamics would be with other large cloud platforms like Google, Amazon, and Microsoft which are also Nvidia’s customers. Nvidia might just see its own platform more as an educational platform and not necessarily a major revenue contributor like AWS long-term.

Currently, there are two potential competitors of Nvidia, one is AMD, but AMD is still struggling to come up with a new GPU to compete. Then there is a ASIC-platform, but most of them are still under development (Intel’s Nervanna) or proprietary (Google’s TPU). So virtually Nvidia is monopolizing the deep learning computing platform.

In this issue, we further analyze on Nvidia’s training plan, the new V100, new partners on Drive PX and its Cloud move. We also cover Medical Imagenet and other news.

As always, if you like our letters, please subscribe and forward it to your colleagues!

Edit at 20170514: Peter Morgan is kind enough to correct us – both Nervanna and TPU are based on ASIC, rather than FPGA. We have corrected the web version since.

Artificial Intelligence and Deep Learning Weekly

News





Blog Posts


Open Source



 

Categories
Uncategorized

AIDL Weekly Issue #12 – Lyrebird, Recursion Pharmaceuticals, and Campas

Editorial

Thoughts From Your Humble Curators

We start to see how machine intelligence can be applied in controversial manner, two related pieces this week:

  • Lyrebird – which astounded us by not only mimicking multiple politicians, but they claim only one minute of training data is enough.
  • Campas – which provides sentence judgement based on software.

We also discuss Recursion Pharmaceuticals and what makes deep learning particularly useful in the company.

As always, if you like our newsletter, remember to subscribe and forward to your colleagues!

Artificial Intelligence and Deep Learning Weekly

News




Blog Posts



Open Source

Video

Member’s Question

Difference Between ML Engineer and Data Scientist?

Q: (From Gautam Karmaker) Guys, what is the difference between ML engineer and a data scientist? How they work together? How their work activity differ? Can you walk through with an use case example?”

A: (From Arthur, redacted)

“Generally, it is hard to decide what a title means unless you know about the nature of the job, usually it is described in the job description. But you can asked what are these terms usually imply. So here is my take:

ML vs data: Usually there is the part of testing/integrating an algorithm and the part of analyzing the data. It’s hard to say how much the proportion on both sides for each job. But high dimensional data is more refrained form simple exploratory analysis. So usually people would use the term “ML” more, which mostly means running/tuning an algorithm. But if you are looking at table-based data, then it’s like to be “data” type of job. IMO, that means at least 40% of your job would be manually looking at trends yourself.

Engineer vs scientist: In larger organization, there is usually a difference between the one who come up with the mathematical model (scientist) vs the one who control the production platform (engineer). e.g. If you are solving a prediction problem, usually scientist is the one who train, say the regression models, but the engineer is the guy who turn your model to create the production system. So you can think of them as the “R” and the “D” in the organization.

Both scientist and engineer are career tracks, and they are equally important. So you would find a lot of companies would have “junior”, “senior”, “principal”, “director”, “VP” prefixed the both track of the titles.

You will sometimes see terms such as programmer or architect replacing “engineer”/”scientist”. Programmer implies their job is more coding-related, i.e. the one who actual write code. Architect is rare, they usually oversee big picture issues among programmers, or act as a balance between R&D organizations.”

Artificial Intelligence and Deep Learning Weekly

About Us

This newsletter is published by Waikit Lau and Arthur Chan. We also run Facebook’s most active A.I. group with 19,000+ members and host a weekly “office hour” on YouTube.

Share on Twitter | Share on Linkedin

Artificial Intelligence and Deep Learning Weekly

©2017-2019 Artificial Intelligence and Deep Learning Weekly
| Sponsorship

 

Categories
Uncategorized

AIDL Weekly Issue 11 – Groq: A Company No One is Talking About

Editorial

Thoughts From Your Humble Curators

Perhaps the biggest news last week is about Groq, a company started by Google’s ex-employees who work on Tensor Processing Unit (TPU). We talk about the company and its current principals.

Of course, ICLR 2017 also held last week. We have two links this issue focused on the conference.

Other than Groq and ICLR 2017, we also cover:

  1. Notes on Stanford cs228n, a Bayesian network class,
  2. a note from Athelas’ Dhruv Parthasarathy on image segmentation,
  3. another criticism of Neuralink.

As always, if you like AIDL Weekly, don’t forget to subscribe and forward to your colleagues!

Artificial Intelligence and Deep Learning Weekly

News


Blog Posts







Video

 

Categories
Uncategorized

AIDL Weekly Issue 10 – F8, Brain-Computer Interface, Apple’s SDC Permit

Editorial

Thoughts From Your Humble Curators

The next big platform everyone will be fighting over is your mind. Check out Elon Musk’s Neuralink and Facebook’s brain-typing and skin-hearing.

Last week also featured F8, which happened on April 18th and 19th. F8 gave us another week filled with some far-out news: Augmented reality? Caffe2.ai? Brain computer interface? Check, check and check. We have 4 items in this issue which cover all these cool stuffs.

We also had a very interesting live-streamed office hour with Sumit Gupta, VP of HPC, AI and Machine Learning at IBM. We went in-depth into what Sumit thinks are bottlenecks in Deep Learning today and other topics. Check out the video below.

Other than F8 and the IBM interview, we also cover:

  • Development of SDC, by Apple and Baidu,
  • Neuralink and its criticism.

As always, if you like our newsletter, subscribe and forward it to your colleagues!

Artificial Intelligence and Deep Learning Weekly

News





Blog Posts


Open Source

Jobs

Video


 

Categories
Uncategorized

AIDL Weekly Issue 9 – Titan Xp or Not? Federated Learning and Last Battle of AlphaGo(?)

Editorial

Thoughts From Your Humble Curators

This week we focus on several note-worthy developments of the week:

  • Federated Learning from Google, what is its impact?
  • Titan Xp, should you buy it or not?
  • AlphaGo vs Ke Jie, is it AlphaGo final battle against humans?
  • Hinton’s NNML class, is it still relevant or not? Should you take it?

As always, if you like our newsletter, subscribe and forward it to your colleagues!

Artificial Intelligence and Deep Learning Weekly

News



Blog Posts





Open Source


Video


About Us

This newsletter is published by Waikit Lau and Arthur Chan. We also run Facebook’s most active A.I. group with 16,000+ members and host a weekly “office hour” on YouTube.

Categories
Uncategorized

AIDL Weekly Issue 8 – Google’s TPU, cs224n 2017, April Fools’ Jokes Roundups Apr 7th 2017

Thoughts From Your Humble Curators

A number of very interesting developments this past week:

  • Google’s TPU,
  • The Vector Institute,
  • Newly released cs224n 2017 videos,
  • CNTK 2.0,
  • BEGAN

Last Saturday was April Fools Day, so we round up the best jokes and pranks about AI. Did you fall for any of them? Some of them, like OpenAI’s spam detection are fairly sophisticated.

As always if you like our newsletter, share it with your friends and colleagues. If you haven’t done it yet, don’t forget to subscribe!

Artificial Intelligence and Deep Learning Weekly


Corrections on Issue #7

In the email edition of Issue 7, we erroneously reported that an autonomous vehicle was involved in a fatal accident. It turns out that there were no serious injuries resulted. We promptly corrected the web version and post correction notice at AIDL. We apologize for causing any misunderstanding.

Artificial Intelligence and Deep Learning Weekly



April Fools’ Jokes on AI and Deep Learning – 2017

It’s almost a week after April Fools, have you fallen for any of the following pranks? Here are some of the best April Fools’ jokes we gathered this year:

The one which we felt confused about: OpenAI’s result on spam detection, because the claim on using simulation to improve real-life training is possible. But the “future plan” on “phishing” and “adversarial spam” give it out. 🙂

(Photo Credit: Open AI)

Artificial Intelligence and Deep Learning Weekly

News




Blog Posts



Open Source

Video


Member’s Question

Should You Learn Lisp?

Q: [As subject]

A: “Learning programming languages, like human languages, or generally different skills, is a way to enlighten you. LISP is a cool language because it does things differently. So sure, in that sense, Lisp may be worth your time.

On the other hand, if you do want to learn modern-day A.I. though, perhaps probability and statistics are the first “language” you want to learn well. As one member, Ernest Szeto said, nowadays A.I. usually use at least some probability-based logic. And if you think probability and statistics as a language, they are fairly difficult to learn on their own.

And yes, at AIDL, we recommend python as the first language, because it allows you to use several stacks in deep learning. You can also use R and java, but notice that there will be a gap between your work and what many people are doing.”

First publish as blog message:

Should You Learn Lisp?

Artificial Intelligence and Deep Learning Weekly

Categories
Uncategorized

AIDL Weekly Issue 7 – The Last Imagenet, OpenAI’s Evolution Strategy and AI Misinformation Epidemic

Thoughts From Your Humble Curators

One of us (Waikit) is teaching a class for MIT in Brisbane, Australia. That’s why we have a lighter issue.

An interesting observation – In the MIT Entrepreneurship classes I’m teaching, there are 120 entrepreneurs from 34 countries spanning U.S to Vietnam to Kazakhstan. One of the top topics of interest and discussion was A.I. and Deep Learning. Surprising or not, some of the students were already implementing fairly advanced DL techniques in agriculture, etc. in emerging economies. It is clear that as A.I. democratizes from the ivory towers of Montreal, Stanford, CMU, FB, Google, Microsoft, etc., there will be some very long-tail positive implications in various economies over time. Is A.I. over-hyped? Sure. But people always over-estimate the short-term and under-estimate the long-term.

This week, we cover:

  • The last ImageNet
  • OpenAI’s new results on Evolution Strategy
  • A new and popular Github, photo style transfer

We also incorporate an article from Zachary Lipton, in which he called out the hype of AI and misinformation spread from popular outlets.

If you like our letter, remember to forward to your friends and colleagues! Enjoy!

Artificial Intelligence and Deep Learning Weekly

News





Blog Posts




The Bandwagon (using in the words of Claude Shannon, 1956)

This is an essay modified from Claude Shannon’s “The Bandwagon” about machine learning. I saw it shared by Cheng Soon Ong.

“Machine Learning has, in the last few years, become something of a scientific bandwagon. Starting as a technical tool for the computer scientist, it has received an extraordinary amount of publicity in the popular as well as the scientific press. In part, this has been due to connections with such fashionable fields computing machines, cybernetics, and automation; and in part, to the novelty of the subject matter. As a consequence, it has perhaps been ballooned to an importance beyond its actual accomplishments. Our fellow scientists in many different fields, attracted by the fanfare and by the new avenues opened to scientific analysis, are using these ideas in their own problems. Applications are being made to biology, psychology, linguistics, fundamental physics, economics, the theory of organisation, and many others. In short, machine learning is currently partaking of a somewhat heady draught of general popularity.

Although this wave of popularity is certainly pleasant and exciting for those of us working in the field, it carries at the same time an element of danger. While we feel that machine learning is indeed a valuable tool in providing fundamental insights into the nature of computing problems and will continue to grow in importance, it is certainly no panacea for the computer scientist or, a fortiori, for anyone else. Seldom do more than a few of natures’ secrets give way at one time. It will be all too easy for our somewhat artificial prosperity to collapse overnight when it is realised that the use of a few exciting words like deep learning, artificial intelligence, data science, do not solve all our problems.

What can be done to inject a note of moderation in this situation? In the first place, workers in other fields should realise that the basic results of the subject are aimed in a very specific direction, a direction that is not necessarily relevant to such fields as psychology, economics, and other social sciences. Indeed, the hard core of machine learning is, essentially, a branch of mathematics and statistics, a strictly deductive system. A thorough understanding of the mathematical foundation and its computing application is surely a prerequisite to other applications. I personally believe that many of the concepts of machine learning will prove useful in these other fields — and, indeed, some results are already quite promising — but the establishing of such applications is not a trivial matter of translating words to a new domain, but rather the slow tedious process of hypothesis and experimental verification. If, for example, the human being acts in some situations like an ideal predictor, this is an experimental and not a mathematical fact, and as such must be tested under a wide variety of experimental situations.

Secondly, we must keep our own house in first class order. The subject of machine learning has certainly been sold, if not oversold. We should now turn our attention to the business of research and development at the highest scientific plane we can maintain. Research rather than exposition is the keynote, and our critical thresholds should be raised. Authors should submit only their best efforts, and these only after careful criticism by themselves and their colleagues. A few first rate research papers are preferable to a large number that are poorly conceived or half-finished. The latter are no credit to their writers and a waste of time to their readers. Only by maintaining a thoroughly scientific attitude can we achieve real progress in machine learning and consolidate our present position.”

Shannon’s original can be found here.

Artificial Intelligence and Deep Learning Weekly

Open Source

Member’s Question

Some Tips on Reading “Deep Learning” By GoodFellow et al

Q: How do you read the book Deep Learning By Ian GoodFellow

It depends on the chapters you are in. The first two parts are better as supplementary material to lectures/courses. For example, if you are reading deep learning and watching all videos from Karpathy’s and Socher’s class, you would learn much more than other students. We think the best lecture to go with is Hinton’s “Neural Network”.

Part 1 tries to power you through the necessary Math. If you never have at least a class of machine learning, those material are woefully inadequate. Consider to study matrix algebra or more importantly matrix differentiation first. (Abadir’s Matrix Algebra is perhaps the most relevant.) Then you will make through the Math more easily. Saying so, Chapter 4’s example on PCA is quite cute. So read them if you are comfortable with the math.

Part 3 is tough, and for the most part it is a reading for researchers in unsupervised learning, which many people believe it is the holy grail of the field. You will need to be comfortable with energy-based model. For that, we suggest you go through Lecture 11 to 15 of Hinton’s deep learning first. If you don’t like unsupervised learning, you could skip Part 3 for now. Reading Part 3 is more about knowing what other people are talking about in unsupervised learning.

While deep learning is a hot field, make sure you don’t abandon other ideas in machine learning. e.g. we find reinforcement learning and genetic algorithm very useful (and fun). Learning theory is deep and can explain certain things we experienced in machine learning. IMO, those topics are at least as interesting as Part 3 of deep learning. (Thanks Richard Green at AIDL for his opinion.)

Artificial Intelligence and Deep Learning Weekly

Categories
Uncategorized

AIDL Weekly Issue #6 – Ng’s Departure, Mask-RCNN and Intel’s AIPG

Thoughts From Your Humble Curators

The biggest AI/DL news last week is definitely Andrew Ng’s departure from Baidu so naturally it is the top news this issue.

Other than Ng’s departure, last week was filled with news on interesting researches and source code:

  • OpenAI’s research on multiple agents led to emergence of a simple language,
  • FAIR’s Kaiming He proposed Mask-RCNN, which shatters previous records,
  • Distill, a new on-line journal for deep learning,
  • Google’s Syntaxnet upgrade,
  • Google’s new skip-thought model.

Enjoy!

Artificial Intelligence and Deep Learning Weekly

News



Blog Posts



Open Source


Video

Book Review

“Deep Learning” by Ian GoodFellow et al

I (Arthur) have some leisure lately to browse “Deep Learning” by Goodfellow for the first time. Since it is known as the bible of deep learning, I decide to write a short afterthought post, they are in point form and not too structured.

  • If you want to learn the zen of deep learning, “Deep Learning” is the book. In a nutshell, “Deep Learning” is an introductory style text book on nearly every contemporary fields in deep learning. It has a thorough chapter covered Backprop, perhaps best introductory material on SGD, computational graph and Convnet. So the book is very suitable for those who want to further their knowledge after going through 4-5 introductory DL classes.
  • Chapter 2 is supposed to go through the basic Math, but it’s unlikely to cover everything the book requires. PRML Chapter 6 seems to be a good preliminary before you start reading the book. If you don’t feel comfortable about matrix calculus, perhaps you want to read “Matrix Algebra” by Abadir as well.
  • There are three parts of the book, Part 1 is all about the basics: math, basic ML, backprop, SGD and such. Part 2 is about how DL is used in real-life applications, Part 3 is about research topics such as E.M. and graphical model in deep learning, or generative models. All three parts deserve your time. The Math and general ML in Part 1 may be better replaced by more technical text such as PRML. But then the rest of the materials are deeper than the popular DL classes. You will also find relevant citations easily.
  • I enjoyed Part 1 and 2 a lot, mostly because they are deeper and fill me with interesting details. What about Part 3? While I don’t quite grok all the Math, Part 3 is strangely inspiring. For example, I notice a comparison of graphical models and NN. There is also how E.M. is used in latent model. Of course, there is an extensive survey on generative models. It covers difficult models such as deep Boltmann machine, spike-and-slab RBM and many variations. Reading Part 3 makes me want to learn classical machinelearning techniques, such as mixture models and graphical models better.
  • So I will say you will enjoy Part 3 if you are 1) a DL researcher in unsupervised learning and generative model or 2) someone wants to squeeze out the last bit of performance through pre-training, 3) someone who want to compare other deep methods such as mixture models or graphical model and NN.

Anyway, that’s what I have now. May be I will summarize in a blog post later on, but enjoy these random thoughts for now.

Original version from my (Arthur’s) blog post.

Artificial Intelligence and Deep Learning Weekly

 

Categories
Uncategorized

AIDL Weekly Issue #5 – Special Issue on Self-Driving Cars

Editorial

Intel/Mobileye big deal, more Waymo/Uber drama, etc. – yet another big week for self-driving cars! It’s not hyperbole to say that self-driving cars represent one of largest market-size application for A.I. The jockeying for positions had been happening for a while and won’t abate anytime soon. Intel largely missed the boat on mobile and is determined not to miss it on A.I. and autonomous vehicles. There’s a subsystem race going on in the h/w and s/w space to solve all the myriad problems.

At the highest level, a successful architecture would need to at least understand:

  • Where am I (car) and where am I going? Need maps, GPS, odometry data.
  • What’s around me based on my sensors? Need car sensors – LIDAR, camera, ultrasound, audio, infrared, etc. Need low-level intelligence / classifiers on each of those signals to identify and make sense of road signs, humans, pets, random objects on the street
  • What’s around me based on external telemetry data? Need other car-related positioning and odometry data, weather data, traffic pattern data
  • How do I make sense of what’s around me, what other objects are doing and whether I’m doing the right actions? A brain that takes internal sensor data and external telemetry data, makes sense of them and outputs an action. This is an oversimplification and is inherently a really tough challenge. There are so many corner and non-corner cases to account for. No company wants to own the first self-driving car that kills a pedestrian. How does the algorithm weigh navigation decisions in an unavoidable accident scenario where you could hit one group of pedestrians or another?
  • How do I train car to be smarter over time? Need phone home feature to a remote human operator if car can’t decide what to do, generating training data
  • Etc.

This isn’t meant to be exhaustive, but as you can see, the moment we start thinking about all the things a human driver does in navigation and in response to other moving blobs on the street, it becomes incredibly hard to create a driving machine replica. We suspect there will be multiple waves of innovation here over time, along the dimensions of better sensors, more types of telemetry data, better cost curve, and better brain.

Artificial Intelligence and Deep Learning Weekly

News




Blog Posts




Open Source


Video

Categories
Uncategorized

AIDL Weekly Issue 4: K for Kaggle, Jetson TX2 and DeepStack

Thoughts From Your Humble Curators

Three big news last week:

  1. Google acquired Kaggle
  2. Jetson TX2 was out,
  3. Just like its rival Libratus, DeepStack made headlines for beating human poker pros.

In this Editorial though, we want to bring to your attention is this little paper titled “Stopping GAN Violence: Generative Unadversarial Networks”. After 1 minute of reading, you would quickly notice that it is a fake paper. But to our dismay, there are newsletters just treat the paper as a serious one. It’s obvious that the “editors” hadn’t really read the original paper.

It is another proof point that the current deep learning space is a over-hyped. Similar happened to Rocket AI). You can get a chuckle out of it but if over-done, it could also over-correct when expectations aren’t met.

Perhaps more importantly, as a community we should spend more conscious effort to fact-check and research a source before we share. We at AIDL Weekly, follow this philosophy religiously and all sources we include are carefully checked – that’s why our newsletter stands out in the crowd of AI/ML/DL newsletters.

If you like what we are doing, check out our FB group, our YouTube channel.

And of course, please share this newsletter with friends so they can subscribe to this newsletter.

Artificial Intelligence and Deep Learning Weekly



Blog Posts



Open Source


Video

Member’s Question

Question from an AIDL Member

Q. (Rephrases from a question asked by Flávio Schuindt) I’ve been studying classification problems with deep learning and now I can understand quite well it. Activation functions, regularizeres, cost functions, etc. Now, I think its time to step forward. What I am really trying to do now is enter in the deep learning image segmentation world. It’s a more complicated problem than classification (object occlusion, lightning variations, etc). My first question is: How can I approach this king of problem? […]

A. You do hit one of the toughest (but hot) problem in deep-learning-based image processing. Many people confuse problems such as image detection/segmentation with image classification. Here are some useful notes.

  1. First of all, have you watched Karpathy’s 2016 cs231n‘s lecture 8 and 13? Those lectures should be your starting points to work on segmentation. Notice that image localization/detection/ segmentation are 3 different things. Localization and detection find bounding boxes and their techniques/concepts can be helpful on “instance segmentation”. “Semantic segmentation” requires downsampling/upsampling architecture. (see below.)
  2. Is your problem more a “semantic segmentation” problem of “instance segmentation” problem? (See cs231n’s lecture 13) The former comes up with regions of different meaning, the latter comes up with instances.
  3. Are you identifying something which always appear? If that’s the case you don’t have to use flunky detection technique, treat it as a localization problem and you can solve by Backprop a simple loss function (as described in cs231n lecture 8). If it might or might not appear, then a detection-type of pipeline might be necessary.
  4. If you do need to use detection-type of pipeline. Does standard segment proposal techniques work for your domain? This is crucial, because at least the beginning of your segmentation research, you have to do find segment proposals.
  5. Lastly if you decide this is really a semantic segmentation problem, then most likely your major task is to adopt an existing pre-train network. Very likely your goal is to transfer learning. Of course check out my point 2 and see if this is really the case.

Artificial Intelligence and Deep Learning Weekly