fix:make zerocolor edge
This commit is contained in:
@@ -128,7 +128,7 @@ enum VideoCaptureAPIs {
|
||||
CAP_INTEL_MFX = 2300, //!< Intel MediaSDK
|
||||
CAP_XINE = 2400, //!< XINE engine (Linux)
|
||||
CAP_UEYE = 2500, //!< uEye Camera API
|
||||
CAP_OBSENSOR = 2600, //!< For Orbbec 3D-Sensor device/module (Astra+, Femto, Astra2, Gemini2, Gemini2L, Gemini2XL, Gemini330, Femto Mega) attention: Astra2 cameras currently only support Windows and Linux kernel versions no higher than 4.15, and higher versions of Linux kernel may have exceptions.
|
||||
CAP_OBSENSOR = 2600, //!< For Orbbec 3D-Sensor device/module (Astra+, Femto, Astra2, Gemini2, Gemini2L, Gemini2XL, Femto Mega) attention: Astra2 cameras currently only support Windows and Linux kernel versions no higher than 4.15, and higher versions of Linux kernel may have exceptions.
|
||||
};
|
||||
|
||||
|
||||
@@ -156,7 +156,7 @@ enum VideoCaptureProperties {
|
||||
CAP_PROP_HUE =13, //!< Hue of the image (only for cameras).
|
||||
CAP_PROP_GAIN =14, //!< Gain of the image (only for those cameras that support).
|
||||
CAP_PROP_EXPOSURE =15, //!< Exposure (only for those cameras that support).
|
||||
CAP_PROP_CONVERT_RGB =16, //!< Boolean flags indicating whether images should be converted to BGR. <br/>
|
||||
CAP_PROP_CONVERT_RGB =16, //!< Boolean flags indicating whether images should be converted to RGB. <br/>
|
||||
//!< *GStreamer note*: The flag is ignored in case if custom pipeline is used. It's user responsibility to interpret pipeline output.
|
||||
CAP_PROP_WHITE_BALANCE_BLUE_U =17, //!< Currently unsupported.
|
||||
CAP_PROP_RECTIFICATION =18, //!< Rectification flag for stereo cameras (note: only supported by DC1394 v 2.x backend currently).
|
||||
@@ -168,7 +168,7 @@ enum VideoCaptureProperties {
|
||||
CAP_PROP_TRIGGER =24,
|
||||
CAP_PROP_TRIGGER_DELAY =25,
|
||||
CAP_PROP_WHITE_BALANCE_RED_V =26,
|
||||
CAP_PROP_ZOOM =27, //!< Android: May switch physical cameras/lenses. Factor and range are hardware-dependent.
|
||||
CAP_PROP_ZOOM =27,
|
||||
CAP_PROP_FOCUS =28,
|
||||
CAP_PROP_GUID =29,
|
||||
CAP_PROP_ISO_SPEED =30,
|
||||
@@ -211,8 +211,6 @@ enum VideoCaptureProperties {
|
||||
CAP_PROP_CODEC_EXTRADATA_INDEX = 68, //!< Positive index indicates that returning extra data is supported by the video back end. This can be retrieved as cap.retrieve(data, <returned index>). E.g. When reading from a h264 encoded RTSP stream, the FFmpeg backend could return the SPS and/or PPS if available (if sent in reply to a DESCRIBE request), from calls to cap.retrieve(data, <returned index>).
|
||||
CAP_PROP_FRAME_TYPE = 69, //!< (read-only) FFmpeg back-end only - Frame type ascii code (73 = 'I', 80 = 'P', 66 = 'B' or 63 = '?' if unknown) of the most recently read frame.
|
||||
CAP_PROP_N_THREADS = 70, //!< (**open-only**) Set the maximum number of threads to use. Use 0 to use as many threads as CPU cores (applicable for FFmpeg back-end only).
|
||||
CAP_PROP_PTS = 71, //!< (read-only) FFmpeg back-end only - presentation timestamp of the most recently read frame using the FPS time base. e.g. fps = 25, VideoCapture::get(\ref CAP_PROP_PTS) = 3, presentation time = 3/25 seconds.
|
||||
CAP_PROP_DTS_DELAY = 72, //!< (read-only) FFmpeg back-end only - maximum difference between presentation (pts) and decompression timestamps (dts) using FPS time base. e.g. delay is maximum when frame_num = 0, if true, VideoCapture::get(\ref CAP_PROP_PTS) = 0 and VideoCapture::get(\ref CAP_PROP_DTS_DELAY) = 2, dts = -2. Non zero values usually imply the stream is encoded using B-frames which are not decoded in presentation order.
|
||||
#ifndef CV_DOXYGEN
|
||||
CV__CAP_PROP_LATEST
|
||||
#endif
|
||||
@@ -232,10 +230,8 @@ enum VideoWriterProperties {
|
||||
VIDEOWRITER_PROP_HW_DEVICE = 7, //!< (**open-only**) Hardware device index (select GPU if multiple available). Device enumeration is acceleration type specific.
|
||||
VIDEOWRITER_PROP_HW_ACCELERATION_USE_OPENCL= 8, //!< (**open-only**) If non-zero, create new OpenCL context and bind it to current thread. The OpenCL context created with Video Acceleration context attached it (if not attached yet) for optimized GPU data copy between cv::UMat and HW accelerated encoder.
|
||||
VIDEOWRITER_PROP_RAW_VIDEO = 9, //!< (**open-only**) Set to non-zero to enable encapsulation of an encoded raw video stream. Each raw encoded video frame should be passed to VideoWriter::write() as single row or column of a \ref CV_8UC1 Mat. \note If the key frame interval is not 1 then it must be manually specified by the user. This can either be performed during initialization passing \ref VIDEOWRITER_PROP_KEY_INTERVAL as one of the extra encoder params to \ref VideoWriter::VideoWriter(const String &, int, double, const Size &, const std::vector< int > ¶ms) or afterwards by setting the \ref VIDEOWRITER_PROP_KEY_FLAG with \ref VideoWriter::set() before writing each frame. FFMpeg backend only.
|
||||
VIDEOWRITER_PROP_KEY_INTERVAL = 10, //!< (**open-only**) Set the key frame interval using raw video encapsulation (\ref VIDEOWRITER_PROP_RAW_VIDEO != 0). Defaults to 1 when not set. FFmpeg back-end only.
|
||||
VIDEOWRITER_PROP_KEY_FLAG = 11, //!< Set to non-zero to signal that the following frames are key frames or zero if not, when encapsulating raw video (\ref VIDEOWRITER_PROP_RAW_VIDEO != 0). FFmpeg back-end only.
|
||||
VIDEOWRITER_PROP_PTS = 12, //!< Specifies the frame presentation timestamp for each frame using the FPS time base. This property is **only** necessary when encapsulating **externally** encoded video where the decoding order differs from the presentation order, such as in GOP patterns with bi-directional B-frames. The value should be provided by your external encoder and for video sources with fixed frame rates it is equivalent to dividing the current frame's presentation time (\ref CAP_PROP_POS_MSEC) by the frame duration (1000.0 / VideoCapture::get(\ref CAP_PROP_FPS)). It can be queried from the resulting encapsulated video file using VideoCapture::get(\ref CAP_PROP_PTS). FFmpeg back-end only.
|
||||
VIDEOWRITER_PROP_DTS_DELAY = 13, //!< Specifies the maximum difference between presentation (pts) and decompression timestamps (dts) using the FPS time base. This property is necessary **only** when encapsulating **externally** encoded video where the decoding order differs from the presentation order, such as in GOP patterns with bi-directional B-frames. The value should be calculated based on the specific GOP pattern used during encoding. For example, in a GOP with presentation order IBP and decoding order IPB, this value would be 1, as the B-frame is the second frame presented but the third to be decoded. It can be queried from the resulting encapsulated video file using VideoCapture::get(\ref CAP_PROP_DTS_DELAY). Non-zero values usually imply the stream is encoded using B-frames. FFmpeg back-end only.
|
||||
VIDEOWRITER_PROP_KEY_INTERVAL = 10, //!< (**open-only**) Set the key frame interval using raw video encapsulation (\ref VIDEOWRITER_PROP_RAW_VIDEO != 0). Defaults to 1 when not set. FFMpeg backend only.
|
||||
VIDEOWRITER_PROP_KEY_FLAG = 11, //!< Set to non-zero to signal that the following frames are key frames or zero if not, when encapsulating raw video (\ref VIDEOWRITER_PROP_RAW_VIDEO != 0). FFMpeg backend only.
|
||||
#ifndef CV_DOXYGEN
|
||||
CV__VIDEOWRITER_PROP_LATEST
|
||||
#endif
|
||||
@@ -268,7 +264,6 @@ enum VideoAccelerationType
|
||||
VIDEO_ACCELERATION_D3D11 = 2, //!< DirectX 11
|
||||
VIDEO_ACCELERATION_VAAPI = 3, //!< VAAPI
|
||||
VIDEO_ACCELERATION_MFX = 4, //!< libmfx (Intel MediaSDK/oneVPL)
|
||||
VIDEO_ACCELERATION_DRM = 5, //!< Raspberry Pi V4
|
||||
};
|
||||
|
||||
//! @} Hardware acceleration support
|
||||
@@ -324,14 +319,14 @@ enum { CAP_PROP_OPENNI_OUTPUT_MODE = 100,
|
||||
#pragma warning( disable: 5054 )
|
||||
#endif
|
||||
//! OpenNI shortcuts
|
||||
enum { CAP_OPENNI_IMAGE_GENERATOR_PRESENT = +CAP_OPENNI_IMAGE_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT,
|
||||
CAP_OPENNI_IMAGE_GENERATOR_OUTPUT_MODE = +CAP_OPENNI_IMAGE_GENERATOR + CAP_PROP_OPENNI_OUTPUT_MODE,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_PRESENT = +CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_BASELINE = +CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_BASELINE,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_FOCAL_LENGTH = +CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_FOCAL_LENGTH,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION = +CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_REGISTRATION,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION_ON = CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION,
|
||||
CAP_OPENNI_IR_GENERATOR_PRESENT = +CAP_OPENNI_IR_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT
|
||||
enum { CAP_OPENNI_IMAGE_GENERATOR_PRESENT = CAP_OPENNI_IMAGE_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT,
|
||||
CAP_OPENNI_IMAGE_GENERATOR_OUTPUT_MODE = CAP_OPENNI_IMAGE_GENERATOR + CAP_PROP_OPENNI_OUTPUT_MODE,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_PRESENT = CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_BASELINE = CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_BASELINE,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_FOCAL_LENGTH = CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_FOCAL_LENGTH,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION = CAP_OPENNI_DEPTH_GENERATOR + CAP_PROP_OPENNI_REGISTRATION,
|
||||
CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION_ON = CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION,
|
||||
CAP_OPENNI_IR_GENERATOR_PRESENT = CAP_OPENNI_IR_GENERATOR + CAP_PROP_OPENNI_GENERATOR_PRESENT,
|
||||
};
|
||||
#ifdef _MSC_VER
|
||||
#pragma warning( pop )
|
||||
@@ -581,18 +576,6 @@ enum { CAP_PROP_ARAVIS_AUTOTRIGGER = 600 //!< Autom
|
||||
|
||||
//! @} ARAVIS
|
||||
|
||||
|
||||
/** @name Android
|
||||
@{
|
||||
*/
|
||||
|
||||
//! Properties of cameras available through NDK Camera API backend
|
||||
enum { CAP_PROP_ANDROID_DEVICE_TORCH = 8001,
|
||||
};
|
||||
|
||||
//! @} Android
|
||||
|
||||
|
||||
/** @name AVFoundation framework for iOS
|
||||
@{
|
||||
*/
|
||||
@@ -714,50 +697,12 @@ enum VideoCaptureOBSensorProperties{
|
||||
CAP_PROP_OBSENSOR_INTRINSIC_FY=26002,
|
||||
CAP_PROP_OBSENSOR_INTRINSIC_CX=26003,
|
||||
CAP_PROP_OBSENSOR_INTRINSIC_CY=26004,
|
||||
CAP_PROP_OBSENSOR_RGB_POS_MSEC=26005,
|
||||
CAP_PROP_OBSENSOR_DEPTH_POS_MSEC=26006,
|
||||
CAP_PROP_OBSENSOR_DEPTH_WIDTH=26007,
|
||||
CAP_PROP_OBSENSOR_DEPTH_HEIGHT=26008,
|
||||
CAP_PROP_OBSENSOR_DEPTH_FPS=26009,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K1=26010,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K2=26011,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K3=26012,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K4=26013,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K5=26014,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_K6=26015,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_P1=26016,
|
||||
CAP_PROP_OBSENSOR_COLOR_DISTORTION_P2=26017
|
||||
};
|
||||
|
||||
//! @} OBSENSOR
|
||||
|
||||
//! @} videoio_flags_others
|
||||
|
||||
/** @brief Read data stream interface
|
||||
*/
|
||||
class CV_EXPORTS_W IStreamReader
|
||||
{
|
||||
public:
|
||||
virtual ~IStreamReader();
|
||||
|
||||
/** @brief Read bytes from stream
|
||||
*
|
||||
* @param buffer already allocated buffer of at least @p size bytes
|
||||
* @param size maximum number of bytes to read
|
||||
*
|
||||
* @return actual number of read bytes
|
||||
*/
|
||||
CV_WRAP virtual long long read(char* buffer, long long size) = 0;
|
||||
|
||||
/** @brief Sets the stream position
|
||||
*
|
||||
* @param offset Seek offset
|
||||
* @param origin SEEK_SET / SEEK_END / SEEK_CUR
|
||||
*
|
||||
* @see fseek
|
||||
*/
|
||||
CV_WRAP virtual long long seek(long long offset, int origin) = 0;
|
||||
};
|
||||
|
||||
class IVideoCapture;
|
||||
//! @cond IGNORED
|
||||
@@ -837,14 +782,6 @@ public:
|
||||
*/
|
||||
CV_WRAP explicit VideoCapture(int index, int apiPreference, const std::vector<int>& params);
|
||||
|
||||
/** @overload
|
||||
@brief Opens a video using data stream.
|
||||
|
||||
The `params` parameter allows to specify extra parameters encoded as pairs `(paramId_1, paramValue_1, paramId_2, paramValue_2, ...)`.
|
||||
See cv::VideoCaptureProperties
|
||||
*/
|
||||
CV_WRAP VideoCapture(const Ptr<IStreamReader>& source, int apiPreference, const std::vector<int>& params);
|
||||
|
||||
/** @brief Default destructor
|
||||
|
||||
The method first calls VideoCapture::release to close the already opened file or camera.
|
||||
@@ -899,19 +836,6 @@ public:
|
||||
*/
|
||||
CV_WRAP virtual bool open(int index, int apiPreference, const std::vector<int>& params);
|
||||
|
||||
/** @brief Opens a video using data stream.
|
||||
|
||||
@overload
|
||||
|
||||
The `params` parameter allows to specify extra parameters encoded as pairs `(paramId_1, paramValue_1, paramId_2, paramValue_2, ...)`.
|
||||
See cv::VideoCaptureProperties
|
||||
|
||||
@return `true` if the file has been successfully opened
|
||||
|
||||
The method first calls VideoCapture::release to close the already opened file or camera.
|
||||
*/
|
||||
CV_WRAP virtual bool open(const Ptr<IStreamReader>& source, int apiPreference, const std::vector<int>& params);
|
||||
|
||||
/** @brief Returns true if video capturing has been initialized already.
|
||||
|
||||
If the previous call to VideoCapture constructor or VideoCapture::open() succeeded, the method returns
|
||||
|
||||
Reference in New Issue
Block a user