Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
AI and Machine Learning – The Basics (steveblank.com)
193 points by sblank on May 18, 2022 | hide | past | favorite | 51 comments


Pretty high-level and broad (which I thought was good). Audience is clearly real-world decision makers as opposed to techies like the HN crowd. Some issues:

1) NLP performance characterized to be better than vision systems. I don't think that is true.

2) Some minor facts are not right. E.g. OpenAI did GPT-3, not Google

3) I expected the set of exemplars for technology to be better researched. Siri and Alexa are NOT recommendation engines. Similarly, lacked the best of class examples on many fronts. This was the biggest issue in the paper.


Help make the paper better. What are some better exemplars for technology?


Sure .. happy to help (and to be clear, I did find the paper to be insightful even as a senior researcher. e.g. I was familiar with OODA but the SPAA was neat! Don't take my comments as too negative :) )

For recommendation systems, the top three examples that come to mind are TikTok, Layer 6 (a Canadian company that TD Bank acquired a few years back) and Netflix.

You may want to add Nerfs to the paper. They are the hot new algorithm out there. I am a scientist at a Canadian research lab and my very smart colleagues tell me it is the next best thing.

Automated vision is far head of NLP IMHO. NLP had it's Imagenet moment only at the advent of BERT (which was cira 2018? or so). Also, Transformers, which BERT and its progeny rely on, are massively compute and data hungry. They are also slow to run on today's chips. In my opinion, reason is that language benchmarks aren't as clearcut as vision. For instance, NLP researchers use BLEU scores, which are a pretty blunt instrument. I'd say NLP is even further behind than speech processing (which is now mostly based on DL). A key person behind Siri is Adam Cheyer btw .. he did Siri and then Bixby. The way these NLP systems work is pretty simple conceptually .. they break the problem to two steps. Intent Identification and then Slot filling. You can use DL for both steps but don't have to. Key issue with NLP systems is they are extremely brittle (a ton of work to customize). Dialog is pretty weak today, and that is partly due to the challenge in training signal.

You say 5000 images per images of a class. I know that was a ballpark but this seemed misleading. There are at least 2 problems I see. First, you need "different" examples .. seeing the same examples (e.g. from different viewpoints) does not help. Second, it really matters what the set of classes your model is trying to discriminate against. E.g. to differentiate apples and bananas, I likely need far fewer than 5000 examples since they are so visually distinct. Imagenet was a seminal moment not just because of the number of examples per class but because of the humongous number of classes (10K+).

For RL and robotics, there have been some neat advances. I was skeptical about RL's utility in practice (due to reasons you point out .. simulations vs. real-world, and especially the issue of faster than real-time) but am seeing it more and more in practice. E.g. 5/6G applications exist.

You may want to add coverage of some important emerging topics: multi-modal (matching vision to text and vice-versa), sensor fusion, student-teacher.

Paper didn't talk about any work from MIT's Han lab or their startup OMNIML? They just launched at the TinyML summit this year and they are hot! Also, TVM tech (startup behind it is called OctoML) is pretty important for on-device AI.

Those were some initial thoughts. If this is useful, can add to it later.


As an ML professor, I agree with all of these comments, especially the bit about Nerfs. Take a look at Waymo's use of the technology: https://waymo.com/research/block-nerf/


Very helpful. I'll update the post tomorrow. thanks.


That is really a good writeup, especially for providing background material for new people in the field (I have been in this field since 1982, experienced AI winters and boom times).

One thing missing in the article is the exponential growth rate of progress. The rate of progress is something I try to explain to non-tech friends. I love seeing difficult problems solved and then simply become new engineering tools to build with.

Transformer models like GPT-3 and Co-Pilot have so quickly transformed my work flow and especially GPT-3 has transformed things that I can do.


This is a good course with examples which can be ran online: https://www.youtube.com/watch?v=_Z9TRANg4c0


Great overview, I like the proper, old school looks of entire site. Also I can recommend this article – https://vas3k.com/blog/machine_learning/ it is on a same subject but has a much more relaxed style.


Is it helpful to anthropomorphise like this:

> An AI can see and understand what it sees. It can identify and detect an object or a feature in an image or video. It can even identify faces.

Why do we switch to this kind of language when we wouldn't for a standard algorithm.

You might say casually "the algorithm has a bug so it doesn't recognize barcodes with a 0 in them" or this door opens when it recognizes a human but it's still very clear that it's not thinking whereas AI people seem to intentionally blur this line to get headlines and attention.

Another example:

> It’s taken decades but as of today, on its simplest implementations, machine learning applications can do some tasks better and/or faster than humans

You wouldn't say this about a pocket calculator, even though it's true. Why say it here?


> You wouldn't say [a pocket calculator can calculate better and faster than humans]

Would. ("Better" is legitimate for "more precisely".)

In fact, to an inquiry, not only «An AI can see and understand what it sees. It can identify and detect» can be read through usual duly interpretation of text as a typical imprecision for «An AI can see and "understand" what it sees: it can identify and detect» (the full stop stands for a colon), where rhetoric 'understand' is explained immediately after, but also the term "understand" is not that necessarily bound to Intelligence: it means "to have entered into a relation, to create a relation" (that 'under-' is a case of 'inter-'), hence "to approach" - "understand" is generally legitimate for progressive (for only progressive) definition of a concept. The use of "understand" for Intelligence must be some elision of "properly, duly understand", proportionally to what is achievable to a human. Which also implies that a more limited entity can "understand" to the best of its nature.

This noted: little problem as per the «barcode», since «it's still very clear».

Surely, if _some_ «people seem to intentionally blur this line», we can censor them.

But surely again, the formulation «whereas AI people seem to intentionally blur this line» is just plain offensive, and the addition «to get headlines and attention» is well over offensive. You have to add the quantifier, "whereas _some_ AI people" - otherwise typicality (beyond statistics: qualification) is implied.


I would still say AI is supremely hyped, and its usage remains in niches. Somehow it still takes thousands of engineers to run Twitter. Not sure how recommendation engine improvements are going to turn the world upside down. I’ve been promised that self driving cars are coming in 6 months for 15 years now, in a perpetual shifting window.


In the course of a normal day, an average person might interact with a dozen different ML-powered apps just using their iPhone.

- Uber/Google Maps/Waze: ETA prediction

- Gmail: Smart Compose & spam filtering

- Instagram/Snapchat/Any camera app: Computer vision

- Siri/Google Assistant: Speech-to-text

- FaceID: Facial recognition

- Facebook/Netflix/All content aggregators: Recommendation engines

- Any banking app: Fraud detection

ML's use is extremely widespread at this point. The above list is just a tiny snapshot. "AI" is term thrown around by marketers and hypemen all the time, no arguments there, but ML's usage is anything but niche these days.


Good list but it is so much more ubiquitous .. any time you pull your camera out to take a picture, there is a ton of AI/Deep Learning based methods being used. Yes, it used to be done with classical techniques a few years back, but DNNs are finding themselves being used in all sort of niche functions.


I mostly agree with you - there is a lot of "something something machine learning!" going on without any concrete use-cases that it enables that wasn't possible before.

I think it will just be another tool available and we'll slowly see it creep into places without making earth shattering improvements. An example that comes to mind is IDE auto-complete that GitHub and others are doing - better than what we had before, but not exactly revolutionary

Apart from within academia and niche places like deepmind, I get the feeling that ML roles and positions will become the new DBA roles of the past - ultimately super-dull jobs where you are just shuffling around storage space, migrating data, or provisioning new tables/models etc without actually adding any business value or getting involved in any of the user-facing work etc.

It is already starting to get heavily commoditized with downloadable pretrained models etc... I see a lot of interns etc who want to come solve cancer with machine learning. I hope that there are not too many youngsters better their career on making it big in machine learning (unless as a researcher) because I genuinely feel like machine learning will just be some library/black box that 99.9% of the time will just be a downloadable pre-trained model that you add like you would if you needed to add OCR support to your product.

I am sure there will be some exciting new things that we'll see of course, but I think they'll be step-wise improvements, rather than huge leaps that open up entirely new worlds of opportunities. E.g. CNNs made image recognition much better (and anyone can now download a really powerful pre-trained model and beat state-of-the-art from just a few years prior etc), but it only made image recognition better - we could do it before, just not as well.


AI has definitely been bastardized and hyped by the marketing departments. Interesting enough recently within the last month IBM has dropped the AI from one of their products.

The sales/pr guy wasn't even sure why or not willing to disclose the why. I wonder if it had to do with unions. However that is total speculations based on an experience with gov't and unions approximately 20 years ago where the unionw was very concerned with computer and software I had written taking over union jobs.


Rule of thumb: if you see a galaxy-brain style graphic next to anything "AI" or "Machine Learning" then the content is probably not worth your time.

This article seems to be an exception to the rule


So Artificial Intelligence is a superset of Machine Learning. What are some AI algorithms that are still in use, that is not Machine Learning? Is there anything?


- Sufficiently complex "if else statements" (Decision tree)

- Expert system / Rule-based system (CLIPS https://clipsrules.net/)

- Knowledge-based system

- Rational agent planning that doesn't involve ML (search, heuristics, MDP, POMDP)

- NLP that doesn't involve ML (Markov model, etc)

If you Google AI taxonomy you can get a good understanding of it.


You covered it in your high level list above (which is pretty good) but I wanted to point out Planners as being an important subfield cutting across different applications.


I think those would fall under the first and last point, depending on implementation.

Most so called algorithms taught in universities are after all some form of basic AI, there's very little conceptual difference between an A* planner and a minimax tree, yet one of these is essentially Stockfish, which is "smarter" than all humans when it comes to chess.


Good list. Also:

- Automated Theorem Proving

- Satisfiability (SAT) and constraint solving


How many of these are in common use today? You rarely hear about things like Expert Systems today. Didn't prolog die in the 1990's?


You don't hear a lot about them nowadays because the hype now is ML, even though things like search or decision tree are probably more widely adopted than ML.

I believe things come and go in cycles, so we might see the rise of non-ML solutions if the current circumstances change. Just like the last time: https://en.wikipedia.org/wiki/AI_winter.


When I took an AI course in 2009 it was largely about game playing algorithms that use graph traversal (possibly guided by heuristics).

Stuff like this: https://en.m.wikipedia.org/wiki/A*_search_algorithm

Such algorithms are still a viable approach in many situations (and I'm sure A* is used in production somewhere today).


In practice these "traditional AI" algorithms are far more common because they're well defined, low cost to implement and run (no gpu instances required lol), and give useful results faster. Proper neural nets are still more buzzwords than workhorses these days, especially when you stray away from image processing where they really shine.

Random forest can predict tabular data extremely well without overfitting, A* is the mathematical best approach to finding the shortest path through a graph, alfa-beta pruning with heuristics consistently beats any known neural net in chess, quicksort will be better than a neural sort because you don't have to first offload data onto a damn gpu, etc.

Even Tesla doesn't give the car's accelerator and steering reigns to the detection neural net outputs (that would be actually insane), instead it lets them extract data from the world into its vector space and then uses deterministic motion planning to determine where the car should drive based on A* GPS routing. That way the system won't do anything undefined and stupid, like neural nets are usually prone to.


Search-based algorithms continue to dominate traditional board games. Stockfish, AlphaGo and friends are hybrid systems combining a game-tree search algorithm with a neural net. The neural net is trained (by self-play) to learn an evaluation function for the game-tree search algorithm (more precisely, the neural nets learn a classifier for board positions as leading to a win, loss or draw). The game-tree search algorithms are alpha-beta minimax in Stockfish and Monte Carlo Tree Search in AlphaGo and family. Far as I know anyway.

DeepMind have downplayed the use of MCTS in their Alpha-x family, to the point of obfuscating the fact that is part of their system at all and have sowed much confusion about this, but their systems ain't going nowhere without good, old-fashion game-tree search.

Stockfish only recently adopted neural nets, btw.


Other replies are missing an explicit call-out to Reinforcement Learning. You can USE ML for RL, but the field itself is considered separate from ML and under AI in general.


RL is generally considered as a type of ML.

Eg Wikipedia: "Reinforcement learning (RL) is an area of machine learning " https://en.m.wikipedia.org/wiki/Reinforcement_learning


The poster probably meant: Reinforcement Learning is not restricted to Neural Networks.


Hmm, I don't see that.

In the spirit of the cutting edge, any chance you could give me a chain-of-reasoning on that inference?


From your link: “Due to its generality, reinforcement learning is studied in many disciplines, such as game theory, control theory, operations research, information theory, simulation-based optimization, multi-agent systems, swarm intelligence, and statistics.”

It all depends on whether you consider the new use as a particular application of a more general thing or as a thing on its own. (But I agree that if you call it with than name it’s not that general.)


The original comment was: "So Artificial Intelligence is a superset of Machine Learning. What are some AI algorithms that are still in use, that is not Machine Learning"

It seems we agree with Wikipedia that ML contains RL?

It's true RL is also studied in other fields.

I struggle to see how that means that RL is a good answer to "What are some AI algorithms that are still in use, that is not Machine Learning".


Does ML contain game theory, control theory, operations research, information theory, simulation-based optimization, multi-agent systems, swarm intelligence, and statistics?

It may be the case if you define ML broadly enough. One may also define RL broadly to refer to things that existed well before ML was a thing (not that I would do it, but one may). I guess that may still be within the AI umbrella, but I’m not sure.


The approach championed by DeepMind is "deep reinforcement learning" because "reinforcement learning" does not automatically imply "deep" (learning).


Reinforcement learning implies learning though. But of course that’s a term more appropriate in the context of “optimizing agents” than in the context of “optimizing prediction models”.


Reinforcement learning is a machine learning approach, there is no serious debate about that. The question is whether it is restricted to neural networks, or not.

For a bit of history on machine learning I recommend Rodney Brooks' seminal series of articles on machine learning, beginnign here:

https://rodneybrooks.com/forai-machine-learning-explained/

The first article in the series, linked above has one section titled "Machine Learning Started with Games". In that section he goes over Arthur Samuel's checkers-playing program that beat a human champion in 1961.

The section also contains Brooks' description of Donald Michie's MENACE, which is widely considered to be one of the first reinforcement learning algorithms. For lack of a computer, it was implemented on a set of match boxes:

In 1960 Surgical Science did not have much pull in getting access to a digital computer. So Donald Michie himself built a machine that could learn to play the game of tic-tac-toe (Noughts and Crosses in British English) from 304 matchboxes, small rectangular boxes which were the containers for matches, and which had an outer cover and a sliding inner box to hold the matches. He put a label on one end of each of these sliding boxes, and carefully filled them with precise numbers of colored beads. With the help of a human operator, mindlessly following some simple rules, he had a machine that could not only play tic-tac-toe but could learn to get better at it.

https://rodneybrooks.com/forai-machine-learning-explained/


> Reinforcement learning is a machine learning approach, there is no serious debate about that. The question is whether it is restricted to neural networks, or not.

The answer to the latter question is obviously “no”. Did anyone argue otherwise? mdp2021 suggested that redytedy may have meant that but what he or she actually wrote is “You can USE ML for RL, but the field itself is considered separate from ML and under AI in general.”


Maybe I'm confused, but I'm replying to feral's OP, where they say "Hmm, I don't see that." in response to mdp2021's comment that "Reinforcement Learning is not restricted to Neural Networks."


That’s how I interpreted the exchange:

mdp2021: “The poster probably meant: Reinforcement Learning is not restricted to Neural Networks.”

feral: “Hmm, I don't see that.” [that RLinrtNN is probably what redytedy meant]

“In the spirit of the cutting edge, any chance you could give me a chain-of-reasoning on that inference?” [the inference that redytedy probably meant RLinrtNN]

I could be wrong.


Yes, maybe I misread that. I read it as feral doubting that Reinforcement Learning is not restricted to neural networks.


I get what you mean. In principle things like policy iteration, value iteration and Q learning are not ML specific.

However, I didn't think of reinforcement learning when compiling that list, because in my experience non-ML RL solutions are rarely better than ML solutions. Happy to be corrected on that front.


Production-rule systems can do inference that isn't from unsupervised neural net training. It's just matching up the rules.

Probably not considered AI these days.


Optical character recognition, applied to scanning documents or automatically reading license plates and postal addresses.


Symbolic regression/program synthesis/genetic programming.


I went to this site and all i saw was a Blank page.


You probably meant, a bit obliquely, to greet Mr. Blank (HN-ID sblank) and subtly thank him for compiling that document and be so kind as to immediately remember this community and submit the information.

https://steveblank.com/about/

We subscribe to the gratitude.


Must be from the school that believes that a "Tabula Rasa" is sufficient as the basics.


Oh, THAT Steve Blank? Not someone I expected to ever write about AI!


Disclaimer - I have started reading this but have not finished it completely. Having said that, it looks very exhaustive and well written and well worth my time to finish it. A bit of feedback to the OP about the page though - the left and right content areas introduce clutter, IMHO, and make it hard to remain focused. But thank you for putting together this resource.



Frankly, I was turned off after the first couple paragraphs because it looked like some sort of government presentation for boomer generals. Will check out further.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: