Giter Club home page Giter Club logo

opencv-machine-learning's Introduction

Machine Learning for OpenCV

Google group DOI

This is the Jupyter notebook version of the following book:


Michael Beyeler
Machine Learning for OpenCV
Intelligent Image Processing with Python

14 July 2017
Packt Publishing Ltd., London, England
Paperback: 382 pages
ISBN 978-178398028-4

The content is available on GitHub. The code is released under the MIT license.

A second edition has now been released (using OpenCV4). The book is also available as a Udemy course.

If you use either book or code in a scholarly publication, please cite as:

M. Beyeler, (2017). Machine Learning for OpenCV. Packt Publishing Ltd., London, England, 380 pages, ISBN 978-178398028-4.

Or use the following bibtex:

@book{MachineLearningOpenCV,
	title = {{Machine Learning for OpenCV}},
	subtitle = {{Intelligent image processing with Python}},
	author = {Michael Beyeler},
	year = {2017},
	pages = {380},
	publisher = {Packt Publishing Ltd.},
	isbn = {978-178398028-4}
}

The book has been referenced more than 45 times in scholarly work (source: Google Scholar). Citing studies include:

  • VS Padilla, RA Ponguillo, AA Abad, LE Salas (2020). Cyber-physical system based on image recognition to improve traffic flow: A case study. International Journal of Electrical and Computer Engineering (IJECE)
  • O Balitskii, V Kolesnikov (2019). Identification of wear products in the automotive tribotechnical system using computer vision methods, artificial intelligence and big data. XIth International Scientific and Practical Conference on Electronics and Information Technologies (ELIT)
  • S Lynch (2018). Image Processing with Python. Dynamical Systems with Applications using Python, Springer.
  • MQG Quiroz (2018). Inductive Machine Learning with Image Processing for Objects Detection of a Robotic Arm with Raspberry PI. International Conference on Technology Trends.

Table of Contents

Preface

Foreword by Ariel Rokem

  1. A Taste of Machine Learning

  2. Working with Data in OpenCV

  3. First Steps in Supervised Learning

  4. Representing Data and Engineering Features

  5. Using Decision Trees to Make a Medical Diagnosis

  6. Detecting Pedestrians with Support Vector Machines

  7. Implementing a Spam Filter with Bayesian Learning

  8. Discovering Hidden Structures with Unsupervised Learning

  9. Using Deep Learning to Classify Handwritten Digits

  10. Combining Different Algorithms Into an Ensemble

  11. Selecting the Right Model with Hyper-Parameter Tuning

  12. Wrapping Up

Running the Code

There are at least two ways you can run the code:

  • deprecated Using Binder (no installation required).
  • Using Jupyter Notebook on your local machine.

The code in this book was tested with Python 3.5, although Python 3.6 and 2.7 should work as well.

Using Binder

Binder allows you to run Jupyter notebooks in an interactive Docker container. No installation required!

Launch the project: mbeyeler/opencv-machine-learning

Using Jupyter Notebook

You basically want to follow the installation instructions in Chapter 1 of the book.

In short:

  1. Download and install Python Anaconda. On Unix, when asked if the Anaconda path should be added to your PATH variable, choose yes. Then either open a new terminal or run $ source ~/.bashrc.

  2. Fork and clone the GitHub repo:

    • Click the Fork button in the top-right corner of this page.
    • Clone the repo, where YourUsername is your actual GitHub user name:
    $ git clone https://github.com/YourUsername/opencv-machine-learning
    $ cd opencv-machine-learning
    
    • Add the following to your remotes:
    $ git remote add upstream https://github.com/mbeyeler/opencv-machine-learning
    
  3. Add Conda-Forge to your trusted channels (to simplify installation of OpenCV on Windows platforms):

    $ conda config --add channels conda-forge
    
  4. Create a conda environment for Python 3 with all required packages:

    $ conda create -n Python3 python=3.6 --file requirements.txt
    
  5. Activate the conda environment. On Linux / Mac OS X:

    $ source activate Python3
    

    On Windows:

    $ activate Python3
    

    You can learn more about conda environments in the Managing Environments section of the conda documentation.

  6. Launch Jupyter notebook:

    $ jupyter notebook
    

    This will open up a browser window in your current directory. Navigate to the folder opencv-machine-learning. The README file has a table of contents. Else navigate to the notebooks folder, click on the notebook of your choice, and select Kernel > Restart & Run All from the top menu.

Getting the latest code

If you followed the instructions above and:

  • forked the repo,
  • cloned the repo,
  • added the upstream remote repository,

then you can always grab the latest changes by running a git pull:

$ cd opencv-machine-learning
$ git pull upstream master

Errata

The following errata have been reported that apply to the print version of the book. Some of these are typos, others are bugs in the code. Please note that all known bugs have been fixed in the code of this repository.

  • p.32: Out[15] should read '3' instead of 'int_arr[3]'.
  • p.32: Out[22] should read array([9, 8, 7, 6, 5, 4, 3, 2, 1, 0]) instead of array([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]).
  • p.33: In the sentence: "Here, the first dimension defines the color channel...", the order of color channels should read "blue, green, and red in OpenCV" instead of "red, green, blue, green, and red".
  • p.36: The range of x values should read "0 <= x <= 10" instead of "0 <= x < 10", since np.linspace by default includes the endpoint.
  • p.51: In [15] shoud read precision = true_positive / (true_positive + false_positive) instead of precision = true_positive / (true_positive + true_negative).
  • p.51: Out[15] should read 0.2 instead of 1.0.
  • p.72: In [6] should read ridgereg = linear_model.Ridge() instead of ridgereg = linear_model.RidgeRegression().
  • p.85: The first line of In [8] should read min_max_scaler = preprocessing.MinMaxScaler(feature_range=(-10,10)) instead of min_max_scaler = preprocessing.MinMaxScaler(feature_range (-10,10)).
  • p.91: The last paragraph should read We also specify an empty array, np.array([]), for the mean argument, which tells OpenCV to compute the mean from the data: instead of We also specify an empty array, np.array([]), for the mask argument, which tells OpenCV to use all data points in the feature matrix:.
  • p.112: In [3] should read vec.get_feature_names()[:5] instead of function:vec.get_feature_names()[:5].
  • p.120: In [16] should read dtree = cv2.ml.DTrees_create() instead of dtree = cv2.ml.dtree_create().
  • p.122: In [26] should read with open("tree.dot", 'w'): f = tree.export_graphviz(dtc, out_file=f, feature_names=vec.get_feature_names(), class_names=['A', 'B', 'C', 'D']) instead of with open("tree.dot", 'w'): f = tree.export_graphviz(clf, out_file=f). Also, the second line should be indented.
  • p.147: The first occurrences of X_hypo = np.c_[xx.ravel().astype(np.float32), yy.ravel().astype(np.float32)] and _, zz = svm.predict(X_hypo) should be removed, as they mistakenly appear twice.
  • p.193: In [28] is missing from sklearn import metrics.
  • p.197: The sentence right below In [3] should read "Then we can pass the preceding data matrix (X) to cv2.kmeans", not cv2.means.
  • p.201: Indentation in bullet points 2-4 are wrong. Please refer to the Jupyter notebook for the correct indentation.
  • p.228: The last sentence in the middle paragraph should read "[...] thus hopefully classifying the sample as y_{hat}=+1" instead of "[...] thus hopefully classifying the sample as y_{hat}=-1".
  • p.230: In [2] has wrong indentation: class Perceptron(object) correctly has indentation level 1, but def __init__ should have indentation level 2, and the two commands self.lr = lr; self.n_iter = n_iter should have indentation level 3.
  • p.260: In [5] should read from keras.models import Sequential instead of from keras.model import Sequential.
  • p.260: In [6] should read model.add(Conv2D(n_filters, (kernel_size[0], kernel_size[1]), padding='valid', input_shape=input_shape)) instead of model.add(Convolution2D(n_filters, kernel_size[0], kernel_size[1], border_mode='valid', input_shape=input_shape)).
  • p.260: In [8] should read model.add(Conv2D(n_filters, (kernel_size[0], kernel_size[1]))) instead of model.add(Convolution2D(n_filters, (kernel_size[0], kernel_size[1]))).
  • p.261: In [12] should read model.fit(X_train, Y_train, batch_size=128, epochs=12, verbose=1, validation_data=(X_test, Y_test)) instead of model.fit(X_train, Y_train, batch_size=128, nb_epoch=12, verbose=1, validation_data=(X_test, Y_test)).
  • p.275, in bullet point 2 it should say ret = classifier.predict(X_hypo) instead of zz = classifier.predict(X_hypo); zz = zz.reshape(xx.shape).
  • p.285: plt.imshow(X[i, :].reshape((64, 64)), cmap='gray') should be indented so that it is aligned with the previous line.
  • p.288: In [14] should read _, y_hat = rtree.predict(X_test) instead of _, y_hat = tree.predict(X_test).
  • p.305: The first paragraph should read "...and the remaining folds (1, 2, and 4) for training" instead of "...and the remaining folds (1, 2, and 4) for testing".
  • p.306: In [2] should read from sklearn.model_selection import train_test_split instead of from sklearn.model_selection import model_selection.
  • p.310: In [18] should read knn.train(X_boot, cv2.ml.ROW_SAMPLE, y_boot) instead of knn.train(X_train, cv2.ml.ROW_SAMPLE, y_boot).
  • p.311: In [20] should have a line model.train(X_boot, cv2.ml.ROW_SAMPLE, y_boot) instead of knn.train(X_boot, cv2.ml.ROW_SAMPLE, y_boot), as well as _, y_hat = model.predict(X_oob) instead of _, y_hat = knn.predict(X_oob).
  • p.328: In [5] is missing the statement from sklearn.preprocessing import MinMaxScaler.
  • p.328: In [5] should have a line pipe = Pipeline([("scaler", MinMaxScaler()), ("svm", SVC())]) instead of pipe = Pipeline(["scaler", MinMaxScaler(), ("svm", SVC())]).

Acknowledgment

This book was inspired in many ways by the following authors and their corresponding publications:

These books all come with their own open-source code - check them out when you get a chance!

opencv-machine-learning's People

Contributors

lbcosta avatar mbeyeler avatar sarthaksuri avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

opencv-machine-learning's Issues

Real-time two-way synchronization between .tex and .ipynb files in the format of ultimate publication.

Hi,

I noted that the project is in the same format as the ultimate publicized book. So I want to achieve the following aims based on the project to generate the corresponding .tex source codes for publication with the following requirements:

  1. The .tex files for this project can be edited and managed by LaTeX IDE, say, texstudio.
  2. The .tex files are generated by .ipynb files of this project.
  3. Each time when I edited the .tex files or the corresponding .ipynb notebooks, synchronization will be triggered between them to ensure the changes are applied to both appropriately.

Is it possible for me to achieve this goal for this project?

Any hints will be highly appreciated.

Best regards
HY

Binder problems

Unfortunately your Binder link seems to be misbehaving.

Binder link provided on homepage

UnsatisfiableError: The following specifications were found to be in conflict:
 - opencv=3.1
 - pandas==0.18.0=np110py27_0
Use "conda info <package>" to see the dependencies for each package.

mybinder doesn’t start up

when launching the mybinder website for your code, mybinder fails to create a process giving the following error:
`
Step 36/52 : RUN chown ${NB_USER}:${NB_USER} ${REPO_DIR}
---> Using cache
---> 1012023c6ff0
Step 37/52 : ENV PATH=${HOME}/.local/bin:${REPO_DIR}/.local/bin:${PATH}
---> Using cache
---> 7e259ec8b997
Step 38/52 : ENV CONDA_DEFAULT_ENV=${KERNEL_PYTHON_PREFIX}
---> Using cache
---> c2d4573adb93
Step 39/52 : COPY --chown=1000:1000 src/environment.yml ${REPO_DIR}/environment.yml
---> Using cache
---> 0100c36cfc80
Step 40/52 : USER ${NB_USER}
---> Using cache
---> 573b8dcd3be0
Step 41/52 : RUN TIMEFORMAT='time: %3R' bash -c 'time ${MAMBA_EXE} env update -p ${KERNEL_PYTHON_PREFIX} --file "environment.yml" && time ${MAMBA_EXE} clean --all -f -y && ${MAMBA_EXE} list -p ${KERNEL_PYTHON_PREFIX} '
---> Running in 56cadc022b99

Looking for: ['python==3.6', 'numpy==1.12', 'scipy==0.19.1', 'scikit-learn==0.18.1', 'matplotlib', 'opencv==3.1', 'jupyter==1.0', 'notebook==5.4.1', 'pandas==0.22', 'theano', 'keras==2.1.5', 'mkl-service==1.1.2']

Could not solve for environment specs
Encountered problems while solving:

  • package libnghttp2-1.46.0-ha19adfc_0 requires openssl >=3.0.0,<4.0a0, but none of the providers can be installed

The environment can't be solved, aborting the operation

�[91mtime: 33.055
�[0mRemoving intermediate container 56cadc022b99
The command '/bin/sh -c TIMEFORMAT='time: %3R' bash -c 'time ${MAMBA_EXE} env update -p ${KERNEL_PYTHON_PREFIX} --file "environment.yml" && time ${MAMBA_EXE} clean --all -f -y && ${MAMBA_EXE} list -p ${KERNEL_PYTHON_PREFIX} '' returned a non-zero code: 1
`

a question about total-variation denoising

Hi Beyeler!

I meet a question about the code denoise(img, weight=0.1, eps=1e-3, num_iter_max=200) you posted at Ask Swiss. Sorry, I can't enter the google group because blocked, and even I can't find any contact to you. And lastly, I post the question under the issue. Do you have a communication software such as skype or QQ? My emial: [email protected]. I look forward to hear from you! Thanks!

precision issue

In notebooks/03.01-Measuring-Model-Performance-with-Scoring-Functions.ipynb,
cell 15

precision = np.sum(true_positive) / np.sum(true_positive + true_negative)

which should be

precision = np.sum(true_positive) / np.sum(true_positive + false_positive)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.