Found in 5 comments
olooney · 2018-10-01 · Original thread
Some of the best textbooks in statistics and machine learning:

Applied -------

Hosmer et al., Applied Logistic Regression. An exhaustive guide to the perils and pitfalls of logistic regression. Logistic regression is the power tool of interpretable statistical models, but if you don't understand it, it will take your foot off (concretely, your inferences will be wrong and your peers will laugh at you.) This book is essential. Graduate level, or perhaps advanced undergraduate, intended for STEM and social science grad students.

https://www.amazon.com/Applied-Logistic-Regression-Probabili...

Peter Christen's Data Matching. Record Linkage is a relatively niche concept, so Christen's book has no right to be as good as it is. But it covers every relevant topic in a clear, even-handed way. If you are working on a record linkage system, then there's nothing in this book you can afford not to know. Undergraduate level, but intended for industry practitioners.

https://www.amazon.com/Data-Matching-Techniques-Data-Centric...

Max Kuhn's Applied Predictive Modeling. Even if you don't use R, this is an incredibly good introduction to how predictive modeling is done in practice. Early undergraduate level.

http://appliedpredictivemodeling.com/

Theoretical -----------

The Elements of Statistical Learning. Probably the single most respected book in machine learning. Exhaustive and essential. Advanced undergraduate level.

https://www.amazon.com/Elements-Statistical-Learning-Predict...

Kevin Murphy's Machine Learning: A Probabilistic Perspective. Covers lots of the same ground as Elements but is a little easier. Undergraduate level.

https://www.amazon.com/Machine-Learning-Probabilistic-Perspe...

Taboga's Lectures on Probability Theory and Mathematical Statistics. Has the distinction of being available for free in web-friendly format at https://www.statlect.com/ . Undergraduate level for math majors.

https://www.amazon.com/Lectures-Probability-Theory-Mathemati...

olooney · 2018-09-10 · Original thread
> In Searle’s time, the dominant AI paradigm was GOFAI (Good Old-Fashioned Artificial Intelligence.)

Russel and Norvig's book is probably the best introduction to "old fashioned" AI:

https://www.amazon.com/Artificial-Intelligence-Modern-Approa...

GOFAI may not have lead directly to true AI, but it produced a ton of useful algorithms such as A* and minimax. Although the attention has turned to machine learning algorithms (à la https://www.amazon.com/Elements-Statistical-Learning-Predict...) the hybrid of GOFAI and ML has produced some extraordinary results, such as AlphaZero:

https://deepmind.com/blog/alphago-zero-learning-scratch/

Thanks for the link.

https://www.amazon.com/Elements-Statistical-Learning-Predict...

vs

https://www.amazon.com/Introduction-Statistical-Learning-App...

The former seems to assume prior knowledge and the latter seems more beginner friendly, except it uses R (in case anyone wants the distinction).

Thanks for the link.

https://www.amazon.com/Elements-Statistical-Learning-Predict...

vs

https://www.amazon.com/Introduction-Statistical-Learning-App...

The former seems to assume prior knowledge and the latter seems more beginner friendly, except it uses R (in case anyone wants the distinction).

View this Book on Amazon