Few friends with me did some works together since last October. All of us were looking for jobs in machine learning or deep learning. We all agreed that we need to review some interesting algorithms together. We had a draft of machine learning algorithms (part 1) during this new year:
Click here for a full version: mlrecap.
Also, we are working on part 2; there are some advanced algorithms which you can see from our outline. It is expected to finish around this June.
These slides are suitable for people to review old things. Some details are not included, so do not suggest readers learn some concepts from our slides. If you find mistakes, please leave comments. If you are interested in some particular algorithms, leave comments and we will consider updating our part 2 outline.
Was working on my research with sklearn, but realized that choosing the right evaluation metrics was always a problem to me. If someone asks me ,”does your model performs well?” The first thing in my mind is “accuracy”. Besides the accuracy, there are a lot, depending on your own problem. Continue reading “Is your model good enough? Evaluation metrics in Classification and Regression”
A brief Introduction here. (Wrote a blog about it last year, but do not think it is detailed.)
This blog is learning notes from this video (English slides but Chinese speaker). First a quick introduction on SVM, then the magic of how to solve max/min values. Also, you could find Kernel SVM. Continue reading “Understanding SVM(2)”
After HMMs, let’s work on a Trigram HMM directly on texts.First will introduce the model, then pieces of code for practicing.
But not going to give a full solution as the course is still going every year, find out more in references.
Continue reading “NLP 02: A Trigram Hidden Markov Model (Python)”
Learning notes for Lecture 7 Modeling sequences: A brief overview. by Geoffrey Hinton 
Continue reading “Deep Learning 10: Sequence Modeling”
When you play any games, probably you have strategies or experiences. But you could not deny that some times you need luck, which data scientists would say a “random choice”. Monte Carlo Method provides only an approximate optimizer, thus giving you the luck to win a game.
Continue reading “Lucky or not: Monte Carlo Method”
A multi-layer NN, able to process images and voice signals (2D), keep stability in rotations.
Continue reading “Deep Learning 05: Talk about Convolutional Neural Networks（CNN）”