change links from samples/python2 to samples/python
This commit is contained in:
parent
2ecb48921b
commit
323e24e3ef
@ -16,8 +16,7 @@ intensity value of the pixel. But in two-dimensional histograms, you consider tw
|
|||||||
it is used for finding color histograms where two features are Hue & Saturation values of every
|
it is used for finding color histograms where two features are Hue & Saturation values of every
|
||||||
pixel.
|
pixel.
|
||||||
|
|
||||||
There is a [python sample in the official
|
There is a python sample (samples/python/color_histogram.py) already
|
||||||
samples](https://github.com/Itseez/opencv/blob/master/samples/python2/color_histogram.py) already
|
|
||||||
for finding color histograms. We will try to understand how to create such a color histogram, and it
|
for finding color histograms. We will try to understand how to create such a color histogram, and it
|
||||||
will be useful in understanding further topics like Histogram Back-Projection.
|
will be useful in understanding further topics like Histogram Back-Projection.
|
||||||
|
|
||||||
@ -106,10 +105,11 @@ You can verify it with any image editing tools like GIMP.
|
|||||||
|
|
||||||
### Method 3 : OpenCV sample style !!
|
### Method 3 : OpenCV sample style !!
|
||||||
|
|
||||||
There is a [sample code for color-histogram in OpenCV-Python2
|
There is a sample code for color-histogram in OpenCV-Python2 samples
|
||||||
samples](https://github.com/Itseez/opencv/blob/master/samples/python2/color_histogram.py). If you
|
(samples/python/color_histogram.py).
|
||||||
run the code, you can see the histogram shows the corresponding color also. Or simply it outputs a
|
If you run the code, you can see the histogram shows the corresponding color also.
|
||||||
color coded histogram. Its result is very good (although you need to add extra bunch of lines).
|
Or simply it outputs a color coded histogram.
|
||||||
|
Its result is very good (although you need to add extra bunch of lines).
|
||||||
|
|
||||||
In that code, the author created a color map in HSV. Then converted it into BGR. The resulting
|
In that code, the author created a color map in HSV. Then converted it into BGR. The resulting
|
||||||
histogram image is multiplied with this color map. He also uses some preprocessing steps to remove
|
histogram image is multiplied with this color map. He also uses some preprocessing steps to remove
|
||||||
|
@ -155,8 +155,8 @@ should be due to the sky)
|
|||||||
|
|
||||||
Well, here you adjust the values of histograms along with its bin values to look like x,y
|
Well, here you adjust the values of histograms along with its bin values to look like x,y
|
||||||
coordinates so that you can draw it using cv2.line() or cv2.polyline() function to generate same
|
coordinates so that you can draw it using cv2.line() or cv2.polyline() function to generate same
|
||||||
image as above. This is already available with OpenCV-Python2 official samples. [Check the
|
image as above. This is already available with OpenCV-Python2 official samples. Check the
|
||||||
Code](https://github.com/Itseez/opencv/raw/master/samples/python2/hist.py)
|
code at samples/python/hist.py.
|
||||||
|
|
||||||
Application of Mask
|
Application of Mask
|
||||||
-------------------
|
-------------------
|
||||||
|
@ -13,7 +13,7 @@ OCR of Hand-written Digits
|
|||||||
|
|
||||||
Our goal is to build an application which can read the handwritten digits. For this we need some
|
Our goal is to build an application which can read the handwritten digits. For this we need some
|
||||||
train_data and test_data. OpenCV comes with an image digits.png (in the folder
|
train_data and test_data. OpenCV comes with an image digits.png (in the folder
|
||||||
opencv/samples/python2/data/) which has 5000 handwritten digits (500 for each digit). Each digit is
|
opencv/samples/data/) which has 5000 handwritten digits (500 for each digit). Each digit is
|
||||||
a 20x20 image. So our first step is to split this image into 5000 different digits. For each digit,
|
a 20x20 image. So our first step is to split this image into 5000 different digits. For each digit,
|
||||||
we flatten it into a single row with 400 pixels. That is our feature set, ie intensity values of all
|
we flatten it into a single row with 400 pixels. That is our feature set, ie intensity values of all
|
||||||
pixels. It is the simplest feature set we can create. We use first 250 samples of each digit as
|
pixels. It is the simplest feature set we can create. We use first 250 samples of each digit as
|
||||||
|
@ -81,7 +81,7 @@ Additional Resources
|
|||||||
Exercises
|
Exercises
|
||||||
---------
|
---------
|
||||||
|
|
||||||
-# OpenCV comes with an interactive sample on inpainting, samples/python2/inpaint.py, try it.
|
-# OpenCV comes with an interactive sample on inpainting, samples/python/inpaint.py, try it.
|
||||||
2. A few months ago, I watched a video on [Content-Aware
|
2. A few months ago, I watched a video on [Content-Aware
|
||||||
Fill](http://www.youtube.com/watch?v=ZtoUiplKa2A), an advanced inpainting technique used in
|
Fill](http://www.youtube.com/watch?v=ZtoUiplKa2A), an advanced inpainting technique used in
|
||||||
Adobe Photoshop. On further search, I was able to find that same technique is already there in
|
Adobe Photoshop. On further search, I was able to find that same technique is already there in
|
||||||
|
@ -156,7 +156,7 @@ in image, there is a chance that optical flow finds the next point which may loo
|
|||||||
actually for a robust tracking, corner points should be detected in particular intervals. OpenCV
|
actually for a robust tracking, corner points should be detected in particular intervals. OpenCV
|
||||||
samples comes up with such a sample which finds the feature points at every 5 frames. It also run a
|
samples comes up with such a sample which finds the feature points at every 5 frames. It also run a
|
||||||
backward-check of the optical flow points got to select only good ones. Check
|
backward-check of the optical flow points got to select only good ones. Check
|
||||||
samples/python2/lk_track.py).
|
samples/python/lk_track.py).
|
||||||
|
|
||||||
See the results we got:
|
See the results we got:
|
||||||
|
|
||||||
@ -213,7 +213,7 @@ See the result below:
|
|||||||

|

|
||||||
|
|
||||||
OpenCV comes with a more advanced sample on dense optical flow, please see
|
OpenCV comes with a more advanced sample on dense optical flow, please see
|
||||||
samples/python2/opt_flow.py.
|
samples/python/opt_flow.py.
|
||||||
|
|
||||||
Additional Resources
|
Additional Resources
|
||||||
--------------------
|
--------------------
|
||||||
@ -221,5 +221,5 @@ Additional Resources
|
|||||||
Exercises
|
Exercises
|
||||||
---------
|
---------
|
||||||
|
|
||||||
-# Check the code in samples/python2/lk_track.py. Try to understand the code.
|
-# Check the code in samples/python/lk_track.py. Try to understand the code.
|
||||||
2. Check the code in samples/python2/opt_flow.py. Try to understand the code.
|
2. Check the code in samples/python/opt_flow.py. Try to understand the code.
|
||||||
|
@ -175,7 +175,7 @@ pattern (every view is described by several 3D-2D point correspondences).
|
|||||||
- A calibration example on stereo matching can be found at
|
- A calibration example on stereo matching can be found at
|
||||||
opencv_source_code/samples/cpp/stereo_match.cpp
|
opencv_source_code/samples/cpp/stereo_match.cpp
|
||||||
- (Python) A camera calibration sample can be found at
|
- (Python) A camera calibration sample can be found at
|
||||||
opencv_source_code/samples/python2/calibrate.py
|
opencv_source_code/samples/python/calibrate.py
|
||||||
|
|
||||||
@{
|
@{
|
||||||
@defgroup calib3d_fisheye Fisheye camera model
|
@defgroup calib3d_fisheye Fisheye camera model
|
||||||
@ -553,7 +553,7 @@ projections, as well as the camera matrix and the distortion coefficients.
|
|||||||
|
|
||||||
@note
|
@note
|
||||||
- An example of how to use solvePnP for planar augmented reality can be found at
|
- An example of how to use solvePnP for planar augmented reality can be found at
|
||||||
opencv_source_code/samples/python2/plane_ar.py
|
opencv_source_code/samples/python/plane_ar.py
|
||||||
- If you are using Python:
|
- If you are using Python:
|
||||||
- Numpy array slices won't work as input because solvePnP requires contiguous
|
- Numpy array slices won't work as input because solvePnP requires contiguous
|
||||||
arrays (enforced by the assertion using cv::Mat::checkVector() around line 55 of
|
arrays (enforced by the assertion using cv::Mat::checkVector() around line 55 of
|
||||||
@ -1674,7 +1674,7 @@ check, quadratic interpolation and speckle filtering).
|
|||||||
|
|
||||||
@note
|
@note
|
||||||
- (Python) An example illustrating the use of the StereoSGBM matching algorithm can be found
|
- (Python) An example illustrating the use of the StereoSGBM matching algorithm can be found
|
||||||
at opencv_source_code/samples/python2/stereo_match.py
|
at opencv_source_code/samples/python/stereo_match.py
|
||||||
*/
|
*/
|
||||||
class CV_EXPORTS_W StereoSGBM : public StereoMatcher
|
class CV_EXPORTS_W StereoSGBM : public StereoMatcher
|
||||||
{
|
{
|
||||||
|
@ -2035,9 +2035,9 @@ so you need to "flip" the second convolution operand B vertically and horizontal
|
|||||||
- An example using the discrete fourier transform can be found at
|
- An example using the discrete fourier transform can be found at
|
||||||
opencv_source_code/samples/cpp/dft.cpp
|
opencv_source_code/samples/cpp/dft.cpp
|
||||||
- (Python) An example using the dft functionality to perform Wiener deconvolution can be found
|
- (Python) An example using the dft functionality to perform Wiener deconvolution can be found
|
||||||
at opencv_source/samples/python2/deconvolution.py
|
at opencv_source/samples/python/deconvolution.py
|
||||||
- (Python) An example rearranging the quadrants of a Fourier image can be found at
|
- (Python) An example rearranging the quadrants of a Fourier image can be found at
|
||||||
opencv_source/samples/python2/dft.py
|
opencv_source/samples/python/dft.py
|
||||||
@param src input array that could be real or complex.
|
@param src input array that could be real or complex.
|
||||||
@param dst output array whose size and type depends on the flags .
|
@param dst output array whose size and type depends on the flags .
|
||||||
@param flags transformation flags, representing a combination of the cv::DftFlags
|
@param flags transformation flags, representing a combination of the cv::DftFlags
|
||||||
@ -2848,7 +2848,7 @@ and groups the input samples around the clusters. As an output, \f$\texttt{label
|
|||||||
|
|
||||||
@note
|
@note
|
||||||
- (Python) An example on K-means clustering can be found at
|
- (Python) An example on K-means clustering can be found at
|
||||||
opencv_source_code/samples/python2/kmeans.py
|
opencv_source_code/samples/python/kmeans.py
|
||||||
@param data Data for clustering. An array of N-Dimensional points with float coordinates is needed.
|
@param data Data for clustering. An array of N-Dimensional points with float coordinates is needed.
|
||||||
Examples of this array can be:
|
Examples of this array can be:
|
||||||
- Mat points(count, 2, CV_32F);
|
- Mat points(count, 2, CV_32F);
|
||||||
|
@ -73,7 +73,7 @@ namespace cv { namespace cuda {
|
|||||||
- A CUDA example applying the HOG descriptor for people detection can be found at
|
- A CUDA example applying the HOG descriptor for people detection can be found at
|
||||||
opencv_source_code/samples/gpu/hog.cpp
|
opencv_source_code/samples/gpu/hog.cpp
|
||||||
- (Python) An example applying the HOG descriptor for people detection can be found at
|
- (Python) An example applying the HOG descriptor for people detection can be found at
|
||||||
opencv_source_code/samples/python2/peopledetect.py
|
opencv_source_code/samples/python/peopledetect.py
|
||||||
*/
|
*/
|
||||||
class CV_EXPORTS HOG : public Algorithm
|
class CV_EXPORTS HOG : public Algorithm
|
||||||
{
|
{
|
||||||
|
@ -74,7 +74,7 @@ This section describes approaches based on local 2D features and used to categor
|
|||||||
- A complete Bag-Of-Words sample can be found at
|
- A complete Bag-Of-Words sample can be found at
|
||||||
opencv_source_code/samples/cpp/bagofwords_classification.cpp
|
opencv_source_code/samples/cpp/bagofwords_classification.cpp
|
||||||
- (Python) An example using the features2D framework to perform object categorization can be
|
- (Python) An example using the features2D framework to perform object categorization can be
|
||||||
found at opencv_source_code/samples/python2/find_obj.py
|
found at opencv_source_code/samples/python/find_obj.py
|
||||||
|
|
||||||
@}
|
@}
|
||||||
*/
|
*/
|
||||||
@ -331,7 +331,7 @@ than union-find method; it actually get 1.5~2m/s on my centrino L7200 1.2GHz lap
|
|||||||
than grey image method ( 3~4 times ); the chi_table.h file is taken directly from paper's source
|
than grey image method ( 3~4 times ); the chi_table.h file is taken directly from paper's source
|
||||||
code which is distributed under GPL.
|
code which is distributed under GPL.
|
||||||
|
|
||||||
- (Python) A complete example showing the use of the %MSER detector can be found at samples/python2/mser.py
|
- (Python) A complete example showing the use of the %MSER detector can be found at samples/python/mser.py
|
||||||
*/
|
*/
|
||||||
class CV_EXPORTS_W MSER : public Feature2D
|
class CV_EXPORTS_W MSER : public Feature2D
|
||||||
{
|
{
|
||||||
|
@ -261,7 +261,7 @@ public:
|
|||||||
|
|
||||||
@note
|
@note
|
||||||
- (Python) A face detection example using cascade classifiers can be found at
|
- (Python) A face detection example using cascade classifiers can be found at
|
||||||
opencv_source_code/samples/python2/facedetect.py
|
opencv_source_code/samples/python/facedetect.py
|
||||||
*/
|
*/
|
||||||
CV_WRAP void detectMultiScale( InputArray image,
|
CV_WRAP void detectMultiScale( InputArray image,
|
||||||
CV_OUT std::vector<Rect>& objects,
|
CV_OUT std::vector<Rect>& objects,
|
||||||
|
@ -107,7 +107,7 @@ objects from still images or video. See <http://en.wikipedia.org/wiki/Inpainting
|
|||||||
- An example using the inpainting technique can be found at
|
- An example using the inpainting technique can be found at
|
||||||
opencv_source_code/samples/cpp/inpaint.cpp
|
opencv_source_code/samples/cpp/inpaint.cpp
|
||||||
- (Python) An example using the inpainting technique can be found at
|
- (Python) An example using the inpainting technique can be found at
|
||||||
opencv_source_code/samples/python2/inpaint.py
|
opencv_source_code/samples/python/inpaint.py
|
||||||
*/
|
*/
|
||||||
CV_EXPORTS_W void inpaint( InputArray src, InputArray inpaintMask,
|
CV_EXPORTS_W void inpaint( InputArray src, InputArray inpaintMask,
|
||||||
OutputArray dst, double inpaintRadius, int flags );
|
OutputArray dst, double inpaintRadius, int flags );
|
||||||
|
@ -74,7 +74,7 @@ See the OpenCV sample camshiftdemo.c that tracks colored objects.
|
|||||||
|
|
||||||
@note
|
@note
|
||||||
- (Python) A sample explaining the camshift tracking algorithm can be found at
|
- (Python) A sample explaining the camshift tracking algorithm can be found at
|
||||||
opencv_source_code/samples/python2/camshift.py
|
opencv_source_code/samples/python/camshift.py
|
||||||
*/
|
*/
|
||||||
CV_EXPORTS_W RotatedRect CamShift( InputArray probImage, CV_IN_OUT Rect& window,
|
CV_EXPORTS_W RotatedRect CamShift( InputArray probImage, CV_IN_OUT Rect& window,
|
||||||
TermCriteria criteria );
|
TermCriteria criteria );
|
||||||
@ -166,9 +166,9 @@ The function implements a sparse iterative version of the Lucas-Kanade optical f
|
|||||||
- An example using the Lucas-Kanade optical flow algorithm can be found at
|
- An example using the Lucas-Kanade optical flow algorithm can be found at
|
||||||
opencv_source_code/samples/cpp/lkdemo.cpp
|
opencv_source_code/samples/cpp/lkdemo.cpp
|
||||||
- (Python) An example using the Lucas-Kanade optical flow algorithm can be found at
|
- (Python) An example using the Lucas-Kanade optical flow algorithm can be found at
|
||||||
opencv_source_code/samples/python2/lk_track.py
|
opencv_source_code/samples/python/lk_track.py
|
||||||
- (Python) An example using the Lucas-Kanade tracker for homography matching can be found at
|
- (Python) An example using the Lucas-Kanade tracker for homography matching can be found at
|
||||||
opencv_source_code/samples/python2/lk_homography.py
|
opencv_source_code/samples/python/lk_homography.py
|
||||||
*/
|
*/
|
||||||
CV_EXPORTS_W void calcOpticalFlowPyrLK( InputArray prevImg, InputArray nextImg,
|
CV_EXPORTS_W void calcOpticalFlowPyrLK( InputArray prevImg, InputArray nextImg,
|
||||||
InputArray prevPts, InputOutputArray nextPts,
|
InputArray prevPts, InputOutputArray nextPts,
|
||||||
@ -213,7 +213,7 @@ The function finds an optical flow for each prev pixel using the @cite Farneback
|
|||||||
- An example using the optical flow algorithm described by Gunnar Farneback can be found at
|
- An example using the optical flow algorithm described by Gunnar Farneback can be found at
|
||||||
opencv_source_code/samples/cpp/fback.cpp
|
opencv_source_code/samples/cpp/fback.cpp
|
||||||
- (Python) An example using the optical flow algorithm described by Gunnar Farneback can be
|
- (Python) An example using the optical flow algorithm described by Gunnar Farneback can be
|
||||||
found at opencv_source_code/samples/python2/opt_flow.py
|
found at opencv_source_code/samples/python/opt_flow.py
|
||||||
*/
|
*/
|
||||||
CV_EXPORTS_W void calcOpticalFlowFarneback( InputArray prev, InputArray next, InputOutputArray flow,
|
CV_EXPORTS_W void calcOpticalFlowFarneback( InputArray prev, InputArray next, InputOutputArray flow,
|
||||||
double pyr_scale, int levels, int winsize,
|
double pyr_scale, int levels, int winsize,
|
||||||
|
@ -380,11 +380,11 @@ class can be used: :
|
|||||||
- Another basic video processing sample can be found at
|
- Another basic video processing sample can be found at
|
||||||
opencv_source_code/samples/cpp/video_dmtx.cpp
|
opencv_source_code/samples/cpp/video_dmtx.cpp
|
||||||
- (Python) A basic sample on using the VideoCapture interface can be found at
|
- (Python) A basic sample on using the VideoCapture interface can be found at
|
||||||
opencv_source_code/samples/python2/video.py
|
opencv_source_code/samples/python/video.py
|
||||||
- (Python) Another basic video processing sample can be found at
|
- (Python) Another basic video processing sample can be found at
|
||||||
opencv_source_code/samples/python2/video_dmtx.py
|
opencv_source_code/samples/python/video_dmtx.py
|
||||||
- (Python) A multi threaded video processing sample can be found at
|
- (Python) A multi threaded video processing sample can be found at
|
||||||
opencv_source_code/samples/python2/video_threaded.py
|
opencv_source_code/samples/python/video_threaded.py
|
||||||
*/
|
*/
|
||||||
class CV_EXPORTS_W VideoCapture
|
class CV_EXPORTS_W VideoCapture
|
||||||
{
|
{
|
||||||
|
@ -31,7 +31,7 @@ if(ANDROID AND BUILD_ANDROID_EXAMPLES)
|
|||||||
endif()
|
endif()
|
||||||
|
|
||||||
if(INSTALL_PYTHON_EXAMPLES)
|
if(INSTALL_PYTHON_EXAMPLES)
|
||||||
add_subdirectory(python2)
|
add_subdirectory(python)
|
||||||
endif()
|
endif()
|
||||||
|
|
||||||
#
|
#
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
if(INSTALL_PYTHON_EXAMPLES)
|
if(INSTALL_PYTHON_EXAMPLES)
|
||||||
file(GLOB install_list *.py )
|
file(GLOB install_list *.py )
|
||||||
install(FILES ${install_list}
|
install(FILES ${install_list}
|
||||||
DESTINATION ${OPENCV_SAMPLES_SRC_INSTALL_PATH}/python2
|
DESTINATION ${OPENCV_SAMPLES_SRC_INSTALL_PATH}/python
|
||||||
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ COMPONENT samples)
|
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ COMPONENT samples)
|
||||||
endif()
|
endif()
|
||||||
|
Loading…
x
Reference in New Issue
Block a user