IMAGE TRANSMISSION
This application is a continuation of International Application No. PCT/CN2018/087089, filed May 16, 2018, the entire content of which is incorporated herein by reference. A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever. The present disclosure relates to information technology and, more particularly, to an image transmission scheme, an image encoding method, an image decoding method, an image transmission system, an encoder, a decoder, a transmitting terminal, a receiving terminal, and an unmanned aerial vehicle (UAV). In wireless image/video transmission systems, the transmission latency is a factor affecting smooth image transmission. Therefore, the size of each image frame obtained after compression is required to match with (or be lower than) the current channel capacity, such that a stable transmission latency is ensured and a smooth transmission is achieved. In digital image transmission systems, the compression of an intra-frame (also referred to as I-frame) does not depend on any other frames, and thus the compression rate of the I-frame is low and the size of the encoded I-frame is large. If the I-frame is forced to reduce the size, the quality of the I-frame becomes poorer. Therefore, the problem that an I-frame is difficult to compress and difficult to match with the channel capacity is a long-standing problem in the image transmission field. Hybrid digital-analog (HDA) transmission systems were introduced to tackle the challenge for matching the image frames with the channel capacity and lowering the transmission latency in the wireless transmission field. The quality of the received images changes with the change of the channel capacity in the HDA transmission systems, such that complex coding mode selection does not need to be performed. However, the conventional HDA transmission systems need higher air interface bandwidth and are less efficient than the pure digital image transmission systems. Some experiments have proved that the quality of the received images using the digital transmission is better than the quality of the received images using the conventional HDA transmission under the same bandwidth and channel environment. In accordance with the disclosure, there is provided an image encoding method including hybrid digital-analog (HDA) encoding a first image frame to generate first encoded data including a digital part and an analog part and inter-encoding a second image frame according to the digital part of the first encoded data to generate second encoded data. Also in accordance with the disclosure, there is provided an image decoding method including HDA decoding first encoded data to obtain a first image frame, the first encoded data including a digital part and an analog part and inter-decoding second encoded data according to the digital part of the first encoded data to obtain a second image frame. Also in accordance with the disclosure, there is provided an encoder including a processor and a memory coupled to the processor and storing instructions. The processor is configured to hybrid digital-analog (HDA) encode a first image frame to generate first encoded data including a digital part and an analog part and inter-encode a second image frame using a reference frame reconstructed from the digital part of the first encoded data to generate second encoded data. Also in accordance with the disclosure, there is provided a decoder including a processor and a memory coupled to the processor and storing instructions. The processor is configured to HDA-decode first encoded data to obtain a first image frame, the first encoded data including a digital part and an analog part and inter-decode second encoded data according to the digital part of the first encoded data to obtain a second image frame. Also in accordance with the disclosure, there is provided an unmanned aerial vehicle (UAV) including a fuselage, a propulsion system coupled to the fuselage, an image acquiring device coupled to the fuselage, and a processor. The propulsion system includes one or more propellers, one or more motors, and an electronic governor. The image acquiring device is configured to acquire a first image frame and a second image frame. The processor is configured to encode the image by HDA encoding the first image frame to generate first encoded data including a digital part and an analog part and inter-encoding the second image frame using a reference frame reconstructed from the digital part of the first encoded data to generate second encoded data. Hereinafter, embodiments consistent with the disclosure will be described with reference to the drawings, which are merely examples for illustrative purposes and are not intended to limit the scope of the disclosure. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts. As shown in In some embodiments, the transmitting terminal 110 can be configured to capture images and perform an image encoding method consistent with the disclosure, such as one of the exemplary image encoding methods described below, on the images to generate the encoded data. The receiving terminal 150 can be configured to receive the encoded data and perform an image decoding method consistent with the disclosure, such as one of the exemplary image decoding methods described below, on the encoded data to recover the images. In some embodiments, the transmitting terminal 110 can be integrated in a mobile object, such as an unmanned aerial vehicle (UAV), a driverless car, a mobile robot, a driverless boat, a submarine, a spacecraft, a satellite, or the like. In some other embodiments, the transmitting terminal 110 can be a hosted payload carried by the mobile object that operates independently but may share the power supply of the mobile object. In some embodiments, the receiving terminal 150 can be a remote controller or a terminal device with an application (app) that can control the transmitting terminal 110 or the mobile object in which the transmitting terminal 110 is integrated, such as a smartphone, a tablet, a game device, or the like. In some other embodiments, the receiving terminal 150 can be provided in another mobile object, such as a UAV, a driverless car, a mobile robot, a driverless boat, a submarine, a spacecraft, a satellite, or the like. The receiving terminal 150 and the mobile object can be separate parts or can be integrated together. The transmission channel 130 can include a wireless channel and/or a wired channel. The transmission channel 130 can use any type of physical transmission medium, such as cable (e.g., twisted-pair wire or fiber-optic cable), air, water, space, or any combination of the above media. For example, if the transmitting terminal 110 is integrated in a UAV and the receiving terminal 150 is a remote controller, the data can be transmitted over air. If the transmitting terminal 110 is a hosted payload carried by a commercial satellite and the receiving terminal 150 is integrated in a ground station, the data can be transmitted over space and air. If the transmitting terminal 110 is a hosted payload carried by a submarine and the receiving terminal 150 is integrated in a driverless boat, the data can be transmitted over water. The image capturing device 111 includes an image sensor and a lens or a lens set, and is configured to capture images. The image sensor can be, for example, an opto-electronic sensor, such as a charge-coupled device (CCD) sensor, a complementary metal-oxide-semiconductor (CMOS) sensor, or the like. The image capturing device 111 is further configured to send the captured images to the encoder 113 for encoding. In some embodiments, the image capturing device 111 can include a memory for storing, either temporarily or permanently, the captured images. The encoder 113 is configured to receive the images captured by the image capturing device 111 and encode the images to generate encoded data. The encoder 113 can support any suitable digital coding standard, such as Moving Picture Experts Group (MPEG, e.g., MPEG-1, MPEG-2, or MPEG-4), H.26x (e.g., H.261, H.262, H.263, or H.264), or the like, any suitable analog coding standard, such as SoftCast or the like, and/or any suitable HDA coding standard, such as WSVC or the like. In some embodiments, the encoder 113 can perform HDA encoding on at least one of the images according to any suitable HDA coding standard and perform digital encoding on the other images according to any suitable digital coding standard. That is, the at least one of the images that is subjected to the HDA encoding can be partially subject to digital encoding to obtain the digital part of the encoded data and can be partially subject to analog encoding to obtain the analog part of the encoded data. In some embodiments, digital encoding may include generating a compressed bit stream from an image using quantization and entropy encoding; while analog encoding may include linearly transforming the image without quantization or entropy encoding. In some embodiments, the at least one of the images that are HDA-encoded is an intra-frame (also referred to as I-frame that is encoded based on information only in the image frame itself), and the images other than the I-frame can be inter-encoded. That is, the images other than the I-frame can be inter-frames (also referred to as P-frames that are encoded with reference to information from one or more different image frames). An image frame may refer to a complete image. Hereinafter, the terms “frame”, “image,” and “image frame” are used interchangeably. Referring again to The first transceiver 115 is configured to obtain the encoded data from the encoder 113 and transmit the encoded data to the receiving terminal 150 over the transmission channel 130. In some other embodiments, the first transceiver 115 can be further configured to receive, for example, the feedback information (e.g., channel information) and/or control commands for controlling the transmitting terminal 110, from the receiving terminal 150 over the wireless channel 130. The first transceiver 115 may work in any suitable frequency band, for example, microwave band, millimeter-wave band, centimeter-wave band, optical wave band, or the like. According to the disclosure, the image capturing device 111, the encoder 113, and the first transceiver 115 can be separate devices, or any two or more of them can be integrated in one device. In some embodiments, the image capturing device 111, the encoder 113, and the first transceiver 115 are separate devices that can be connected or coupled to each other. For example, the image capturing device 111 can be a camera, a camcorder, or a smartphone having a camera function. The encoder 113 can be an independent device including a processor and a memory as shown in In some other embodiments, any two of the image capturing device 111, the encoder 113, and the first transceiver 115 can be integrated in a same device. For example, the image capturing device 111 and the encoder 113 may be parts of a same device including a camera, a lens, a processor, and a memory. The processor can be any type of processor and the memory can be any type of memory. The disclosure is not limited here. In this example, the device can further include an electrical interface (either wired or wireless) for coupling to the first transceiver 115. In some other embodiments, the image capturing device 111, the encoder 113, and the first transceiver 115 can be integrated in a same electronic device. For example, the image capturing device 111 may include an image sensor and a lens or a lens set of the electronic device. The encoder 113 may be implemented by a single-chip encoder, a single-chip codec, an image processor, an image processing engine, or the like, which is integrated in the electronic device. The first transceiver 115 may be implemented by an integrated circuit, a chip, or a chipset that is integrated in the electronic device. For example, the electronic device may be a smartphone having a built-in camera and a motherboard that integrates the encoder 113 and the first transceiver 115. The second transceiver 151 is configured to receive the encoded data from the transmitting terminal 110 over the wireless channel 130 and send the encoded data to the decoder 153 for decoding. In some other embodiments, the second transceiver 151 is further configured to transmit, for example, feedback information (e.g., channel information) and/or control commands for controlling the transmitting terminal 110, to the transmitting terminal 110 over the wireless channel 130. The second transceiver 151 can include a transmitter and a receiver. The second transceiver 151 can be configured to have a two-way communications capability. In some embodiments, the transmitter and the receiver may share common circuitry. In some other embodiments, the transmitter and the receiver may be separate parts sharing a single housing. The second transceiver 151 can work in a same frequency band as that used in the first transceiver 115 of the transmitting terminal 110. For example, if the first transceiver 115 uses the microwave band, the second transceiver 151 works in the corresponding microwave band. If the first transceiver 115 uses optical wave band, the second transceiver 151 works in the corresponding optical wave band. The decoder 153 is configured to obtain the encoded data from the second transceiver 151 and decode the encoded data to recover the images captured by the image capturing device 111. The decoder 153 can support any digital coding standard that is employed in the encoder 113, any analog coding standard that is employed in the encoder 113, and/or any HDA coding standard that is employed in the encoder 113. In some embodiments, the at least one of images subject to the HDA encoding, i.e., partial digital encoding and partial analog encoding, in the encoder 113 of the transmitting terminal 110 can be recovered by the decoder 153, according to the HDA coding standard that is employed by the encoder 113 of the transmitting terminal 110. The images subject to the digital encoding in the encoder 113 of the transmitting terminal 110 can be recovered by the decoder 153, according to the digital coding standard that is employed by the encoder 113 of the transmitting terminal 110. In some embodiments, the decoder 153 can decode the I-frames using the corresponding HDA coding standard and decode the P-frames using the corresponding digital coding standard. In some embodiments, the decoder 153 can inter-decode the P-frames using the corresponding digital coding standard. Referring again to In some embodiments, the screen 155 can include a touch panel for receiving a user input. The user can touch the screen 155 with an external object, such as a finger of the user or a stylus. In some embodiments, the user can adjust image parameters, such as brightness, contrast, saturation, and/or the like, by touching the screen 155. For example, the user can scroll vertically on the image to select a parameter, then swipe horizontally to change the value of the parameter. In some embodiments, the user can input the control command for controlling the transmitting terminal 110 by touching the screen 155. For example, the user can input a control command for controlling the image capturing device 111 of the transmitting terminal 110 to start or stop capturing images. As another example, the user can input a control command for selecting the coding technique used in the encoder 113 of the transmitting terminal 110. The screen 155 can also be configured to send the control command inputted by the user to the second transceiver 151, such that the second transceiver 151 can transmit the control command to the transmitting terminal 110. According to the disclosure, the second transceiver 151, the decoder 153, and the screen 155, can be separate devices, or any two or more of them can be integrated in one device. In some embodiments, the second transceiver 151, the decoder 153, and the screen 155 are separate devices that can be connected or coupled to each other. For example, the second transceiver 151 can be an independent device combining transmitter/receiver in a single package. The decoder 153 can be an independent device including the processor and the memory as shown in In some other embodiments, any two of the second transceiver 151, the decoder 153, and the screen 155 can be integrated in a same device. For example, the decoder 153 and the screen 155 may be parts of a same device including a processor, a memory, and a screen. The processor can be any type of processor and the memory can be any type of memory. The disclosure is not limited here. In this example, the device can further include an electrical interface (either wired or wireless) for coupling to the second transceiver 151. In some other embodiments, the second transceiver 151, the decoder 153, and the screen 155 are integrated in a same electronic device. For example, the second transceiver 151 may be implemented by an integrated circuit, a chip, or a chipset that is integrated in the electronic device. The decoder 153 may be implemented by a single-chip decoder, a single-chip codec, an image processor, an image processing engine, or the like, which is integrated in the electronic device. For example, the electronic device may be a tablet having a screen and a motherboard that integrates the second transceiver 151 and the decoder 153. The navigation system 602 can include one or more of motion sensors (e.g., accelerometers), rotation sensors (e.g., gyroscopes), magnetic sensors (magnetometers), or the like. The navigation system 602 can be configured to detect a speed, an acceleration, and/or attitude parameters (such as pitch angle, roll angle, yaw angle, and/or the like) of the UAV 600, attitude parameters of the image acquiring device 604, and/or attitude parameters of the gimbal 605. The navigation system 602 can be provided inside or on the fuselage 601 of the UAV 600. The control system 603 is coupled to the navigation system 602, the electronic governor 613, and the gimbal 605. The control system 603 can be configured to control a flight attitude of the UAV 600 and/or a rotation of the gimbal 604 according the attitude parameters obtained by the navigation system 602. In some embodiments, the control system 603 can be coupled to the image acquiring device 604 and configured to control an attitude, such as a rotation, of the image acquiring device 604. The control system 603 can be provided inside the fuselage 601 of the UAV 600. The image acquiring device 604 is connected to the fuselage 601 of the UAV 600 via the gimbal 605. In some embodiments, the image acquiring device 604 can be directly connected to the fuselage 601 without the need for the gimbal 605. The image acquiring device 604 can be provided below or above the fuselage 601 of the UAV 600. The image acquiring device 604 can include an image sensor and a lens or a lens set. The image acquiring device 604 is configured to capture the images. The image sensor can include, for example, an opto-electronic sensor, such as a charge-coupled device (CCD) sensor, a complementary metal-oxide-semiconductor (CMOS) sensor, or the like. The image acquiring device 604 can rotate along with the rotation of the gimbal 605, such that the image acquiring device 604 can perform a tracking shooting on a target object. In some embodiments, the image acquiring device 604 can include an encoder (not shown in The communication system 606 can include a receiver and/or a transmitter. The receiver can be configured to receive wireless signals 620 transmitted by an antenna 631 of a ground station 632, and the communication system 606 can also send the wireless signals 620 (such as the encoded data, status information of the UAV, or the like) to the ground station 632. The communication system 606 can be similar to the first transceiver 115 described above. The communication system 606 can be provided inside or on the fuselage 601 of the UAV 600. Exemplary image transmission scheme consistent with the disclosure will be described in more detail below. An image transmission scheme consistent with the disclosure can be implemented in an image transmission system consistent with the disclosure, such as the image transmission system 100 described above. As shown in The image frame (denoted as P0 in Other image frames (denoted as P1, P2, P3, . . . in The encoded data IA, ID, PD0, PD1, PD2, PD3, . . . can be transmitted by the transmitting terminal through a transmission channel, such as the transmission channel 130 described above, and received by a receiving terminal consistent with the disclosure, such as the receiving terminal 150 described above. The corresponding received encoded data are denoted as IA′, ID′, PD0′, PD1′, PD2′, PD3′, . . . , respectively in In some embodiments, the received digital part ID′ and the received analog part IA′ received at the receiving terminal can be decoded using the HDA coding standard to obtain a recovered image frame I′ corresponding to the image frame I. That is, the image frame I can be recovered as the recovered image frame I′ by intra-decoding the received digital part ID′ and analog-decoding the received analog part IA′. For example, the low frequency components of the recovered image frame I′ can be obtained from the received digital part ID′ and the high frequency components of the recovered image frame I′ can be obtained from the received analog part IA′. The bitstream PD0′ received immediately after the received digital part ID′ and the received analog part IA′ corresponding to the image frame I can be inter-decoded according to the received digital part ID′ corresponding to the image frame I to obtain a recovered image frame P0′ corresponding to the image frame P0. That is, the bitstream PD0′ can be inter-decoded with reference to a decoded image recovered from the received digital part ID′ corresponding to the image frame I. The other bitstreams PD1′, PD2′, PD3′, . . . received after the bitstream PD0′ and between two I-frames can be inter-decoded using the digital coding standard. In some embodiments, the reference frame for a bitstream other than PD0′ can be a previously decoded frame. In some other embodiments, the reference frame for a bitstream other than PD0′ can include multiple previously decoded frames. The HDA coding standard, the digital coding standard, and/or the analog coding standard used in the receiving terminal can be the same as those used in the transmitting terminal. In the transmitting terminal, the encoding operations of the HDA coding standard, the digital coding standard, and/or the analog coding standard are performed. In the receiving terminal, the decoding operations of the HDA coding standard, the digital coding standard, and/or the analog coding standard are performed. The HDA coding standard can be any type of HDA coding standard; the digital coding standard can be any type of digital coding standard; and the analog coding standard can be any type of analog coding standard. The disclosure is not limited here. The selection of the coding standard can be based on the actual needs. Exemplary image encoding methods consistent with the disclosure will be described in more detail below. An image encoding method consistent with the disclosure can be implemented in a transmitting terminal consistent with the disclosure, such as the transmitting terminal 110 of the image transmission system 100 described above. As shown in Here, a whole frame or a block of the frame, such as a macroblock (MB), a sub-block, or the like can be encoded. The block of the first image frame refers to a portion of the first image frame, which includes a plurality of pixels of the first image frame. As shown in The prediction residual data can be generated by subtracting intra-predicted data of the first image frame from the first image frame. The intra-predicted data of the first image frame can be generated by performing intra-prediction on the first image frame using one of a plurality of intra-prediction modes. Similarly, the intra-predicted data can also be in a 2D form. In some embodiments, the plurality of intra-prediction modes can be those supported by the digital coding standard that is employed. The one of the plurality of intra-prediction modes can be one that is most suitable for the first image frame, which is also referred to as a best intra-prediction mode. For example, the digital coding standard H.264 supports nine intra-prediction modes for luminance 4×4 and 8×8 blocks, including 8 directional modes and one intra direct component (DC) mode that is a non-directional mode. In this situation, the best intra-prediction mode for the first image frame can be selected from all intra-prediction modes supported by H.264 as described above. Any suitable intra-prediction mode selection technique can be used here. For example, a Rate-Distortion Optimization (RDO) technique can be used to select the best intra-prediction mode which has a least rate-distortion (RD) cost. The intra-predicted data can be subtracted from the first image frame to generate the prediction residual data. At 903, the prediction residual data is transformed into transform coefficients. That is, the prediction residual data is transformed from the spatial domain into a representation in the spatial frequency domain for more efficient quantization and data compression. In the spatial frequency domain, the prediction residual data can be expressed in terms of a plurality of frequency-domain components, such as a plurality of sine and/or cosine components. Coefficients associated with the frequency-domain components in the frequency-domain expression are also referred to as the transform coefficients. Similarly, the transform coefficients can also be arranged in a 2D form. Any suitable transform algorithm can be used to obtain the transform coefficients, such as discrete cosine transform (DCT), discrete wavelet transform (DWT), time-frequency analysis, Fourier transform, lapped transform, or the like. For example, in H.264, a residual block can be transformed using a 4×4 or 8×8 integer transform derived from the DCT. At 905, the analog part of the first encoded data is generated according to the transform coefficients corresponding to high frequency components of the first image frame. In some embodiments, the analog part of the first encoded data can include the transform coefficients corresponding to the high frequency components of the first image frame. In some embodiments, the high frequency components that almost do not contribute to information in the first image frame, i.e., the high frequency components that have very small transform coefficients, such as the high frequency components having transform coefficients smaller than a threshold value, can be discarded. That is, the analog part of the first encoded data can exclude the transform coefficients corresponding to the high frequency components that do not contribute to information in the first image frame. For example, zero- and near-zero-value transform components can be discard. That is, the transform coefficients having a zero or near-zero value can be excluded from the analog part of the first encoded data. In some embodiments, nearby transform components can be grouped in one chunk and a decision can be made for all transform components in a chunk. That is, all transform components in one chunk can be retained or discarded together. Making one decision per chunk allows reducing an amount of metadata to a decoder of the receiving terminal, such as the decoder 153 of the receiving terminal 150 described above, for locating, e.g., the discarded transform components. At 907, the transform coefficients corresponding to the low frequency components of the first image frame are quantized to generate quantized transform coefficients. In some embodiments, the transform coefficients corresponding to the low frequency components are divided by a quantization step size (Qstep) to obtain the quantized transform coefficients. A larger value of the quantization step size results in a higher compression at the expense of a poorer image quality. Similarly, the quantized transform coefficients can be also in a 2D form. At 909, the quantized transform coefficients are entropy encoded to generate the digital part of the first encoded data. That is, the quantized transform coefficients are converted into binary codes, i.e., the digital part of the first encoded data. Any suitable entropy encoding technique may be used, such as Huffman coding, Unary coding, Arithmetic coding, Shannon-Fano coding, Elias gamma coding, Tunstall coding, Golomb coding, Ricde coding, Shannon coding, Range encoding, universal coding, exponential-Golomb coding, Fibonacci coding, or the like. In some embodiments, the quantized transform coefficients may be reordered before being subject to the entropy encoding. In some embodiments, the prediction process at 901 can be omitted. That is, the first image frame can be directly transformed to obtain the transform coefficients without prediction. As shown in The first image frame can be divided into a plurality of blocks. The number and size of the blocks can be determined according the actual needs. The amounts of information in the plurality of blocks can be calculated and characterized by, for example, information entropies of the plurality of blocks. Generally, a larger entropy value corresponds to a larger amount of information and a smaller entropy value corresponds to a less amount of information. Any parameters that can reflect the amounts of information in the plurality of blocks can be used here. The disclosure is not limited here. A block can either belong to the high-information portion of the first image frame or the low-information portion of the first image frame, depending on the amount of information of the block. That is, the high-information portion of the first image frame refers to a portion of the first image frame that contains blocks with large amounts of information, while the low-information portion of the first image frame refers to a portion of the first image frame that contains blocks with small amounts of information. For example, a block having the amount of information less than or equal to a threshold can belong to the low-information portion of the first image frame and a block having the amount of information greater than the threshold can belong to the high-information portion of the first image frame. The threshold can be determined according to at least one of a channel bandwidth, a bit rate, or a resolution of the first image frame. At 1040, the digital part of the first encoded data is generated according to the high-information portion of the first image frame. That is, the high-information portion of the first image frame is digital-encoded, e.g., intra-encoded, to generate the digital part of the first encoded data. Digital-encoding, e.g., intra-encoding, the high-information portion of the first image frame can be accomplished according to any suitable digital coding standard, such as MPEG-x (e.g., MPEG-1, MPEG-2, or MPEG-4), H.26x (e.g., H.261, H.262, H.263, or H.264), or another format. The disclosure is not limited here. Intra-encoding the high-information portion of the first image frame can include applying intra-prediction, transform, quantization, and entropy encoding to the high-information portion of the first image frame. In some embodiments, the high-information portion of the first image frame can be processed/encoded block by block. The intra-prediction, transform, quantization, and entropy encoding processes for intra-encoding the high-information portion of the first image frame are similar to those processes in At 1060, the analog part of the first encoded data is generated according to the low-information portion of the first image frame. That is, the low-information portion of the first image frame can be analog-encoded to generate the analog part of the first encoded data. Analog-encoding the low-information portion of the first image frame can be accomplished according to any suitable analog coding standard, such as SoftCast, line-cast, Realcast, or the like. The disclosure is not limited here. In some embodiments, analog-encoding the low-information portion of the first image frame can include generating the analog part of the first encoded data according to the transform coefficients of the low-information portion of the first image frame. Any suitable transform algorithm can be used to obtain the transform coefficients, such as, for example, DCT, DWT, three dimensional DCT (3D-DCT), 2D-DWT+DCT, or the like. For example, in the DCT transform, the low-information portion of the first image frame can be expressed in terms of a plurality DCT components. Coefficients associated with the DCT components can form the analog part of the first encoded data and can be directly transmitted without quantization and entropy encoding. In some embodiments, analog-encoding the low-information portion of the first image frame can also include discarding the frequency components, e.g., the DCT components, that do not contribute to the information in the low-information portion of the first image frame. For example, zero- or near-zero-value DCT components can be discarded. In some embodiments, the nearby DCT components can be grouped in one chunk and a decision can be made for all DCT components in a chunk. That is, all DCT components in the chunk can be retained or discarded together. As noted above, making one decision per chunk allows reducing an amount of metadata to a decoder of the receiving terminal, such as the decoder 153 of receiving terminal 150 described above, for locating, e.g., the discarded DCT components. Referring again to Inter-encoding the second image frame can be accomplished according to any suitable digital coding standard, such as MPEG-x (e.g., MPEG-1, MPEG-2, or MPEG-4), H.26x (e.g., H.261, H.262, H.263, or H.264), or another standard. Inter-encoding the second image frame can include applying inter-prediction, transform, quantization, and entropy encoding to the second image frame. In some embodiments, applying the inter-prediction process on the second image frame can include generating inter-predicted data of the second image frame with reference to the reconstructed frame of the digital part of the first image frame using one of a plurality of inter-prediction modes. In some embodiments, the plurality of inter-prediction modes can be those supported by the digital coding standard that is employed. The one of the plurality of inter-prediction modes can be one that is most suitable for the second image frame, which is also referred to as a best inter-prediction mode. For example, if H.264 is employed, the inter-prediction can use one of a plurality of block sizes, e.g., a size of 16×16, a size of 16×8, a size of 8×16, a size of 8×8, a size of 8×4, a size of 4×8, and a size of 4×4. The inter-prediction in H.264 also includes a block matching process, during which a best matching block is identified as a reference block for the purposes of motion estimation. The best matching block for a block of the second image frame can be a block in the reconstructed frame of the digital part of the first image frame that is most similar to the block of the second image frame. That is, there is a smallest prediction error between the best matching block and the block of the second image frame. Any suitable block matching algorithm can be employed, such as exhaustive search, optimized hierarchical block matching (OHBM), three step search, two dimensional logarithmic search (TDLS), simple and efficient search, four step search, diamond search (DS), adaptive rood pattern search (ARPS), or the like. In this situation, the best inter-prediction mode for the second image frame can be selected from all possible combinations of the inter-prediction modes supported by H.264 as described above. Any suitable inter-prediction mode selection technique can be used here. For example, an RDO technique selects the best inter-prediction mode which has a least RD cost. The inter-predicted data can be subtracted from the second image frame to generate prediction residual data. The transform, quantization, and entropy encoding processes for inter-encoding the second image frame are similar to those described above in connection with Inter-encoding the second image frame can also include generating the reconstructed frame by reconstructing the digital part of the first encoded data as the reference frame for encoding the second image frame. Generating the reconstructed frame of the digital part of the first encoded data can include applying inverse quantization, inverse transform, and inter-prediction on the quantized transform coefficients corresponding to the digital part of the first encoded data. The inverse quantization, inverse transform, and inter-prediction processes are similar to those in the decoding processes described below, and thus detailed description thereof is omitted here. In some embodiments, when the digital part of the first encoded data is generated according to the high-information portion of the first image frame as shown in Referring again to The processes of inter-encoding the third image frame with reference to the second image frame to generate the third encoded data are similar to the processes of inter-encoding the second image frame with reference to the reconstructed frame of the digital part of the first image frame at 804. The detailed description thereof is omitted here. In some embodiments, any image frame after the second image frame and between the first I-frame (i.e., the first image frame) and a second I-frame, i.e., an I-frame following the first I-frame without another I-frame therebetween, can be inter-encoded with reference to a past image frame (a neighboring frame that was obtained before the image frame). Exemplary image decoding methods consistent with the disclosure will be described in more detail below. An image decoding method consistent with the disclosure can be implemented in a receiving terminal consistent with the disclosure, such as the receiving terminal 150 of the image transmission system 100 described above. As shown in Here, a whole frame or a block of the frame, such as a macroblock (MB), a sub-block, or the like can be decoded, corresponding to the encoding process. For example, if a whole frame is encoded all together at once, the whole frame can be decoded all together. On the other hand, if a frame is encoded block by block, the frame can be decoded block by block. As shown in The entropy decoding process can convert the digital part of the first encoded data into the decoded quantized transform coefficients. An entropy decoding technique corresponds to the entropy encoding technique used for generating the digital part of the first image frame can be used. For example, when Huffman coding is employed in the entropy encoding process, Huffman decoding can be used in the entropy decoding process. As another example, when Arithmetic coding is employed in the entropy encoding process, Arithmetic decoding can be used in the entropy decoding process. At 1204, decoded transform coefficients corresponding to the low frequency components are obtained by inversely quantizing the decoded quantized transform coefficients. The decoded quantized transform coefficients can be multiplied by the the quantization step size Qstepto generate the decoded transform coefficients corresponding to the low frequency components. At 1206, decoded transform coefficients corresponding to the high frequency components are obtained according to the analog part of the first encoded data. In some embodiments, the decoded transform coefficients corresponding to the high frequency components are directly included in the analog part of the first encoded data. In some embodiments, the coefficients of the frequency components that were discarded at 905 can be substituted by zero. At 1208, combined decoded transform coefficients are obtained by combining the decoded transform coefficients corresponding to the low frequency components with the decoded transform coefficients corresponding to the high frequency components. For example, the decoded transform coefficients corresponding to the low frequency components can be added to the decoded transform coefficients corresponding to the high frequency components to form the combined decoded transform coefficients. At 1210, the decoded prediction residual data are obtained by inversely transform the combined decoded transform coefficients. An inverse transform algorithm corresponds to the transform algorithm employed for encoding the first image frame can be used here. For example, in H.264, if the 4×4 or 8×8 integer transform derived from the DCT is employed in the transform process, the 4×4 or 8×8 inverse integer transform can be used in the inverse transform process. At 1212, the first image frame is recovered from the prediction residual data. In some embodiments, recovering the first image frame can include obtaining recovered predicted data according to a prediction mode. A prediction mode corresponds to the intra-prediction mode that is employed for intra-encoding the first image frame can be used in obtaining the recovered predicted data. The implementation of the prediction process is similar to the implementation of the intra-prediction process described above at 903. The detailed description thereof is omitted here. The decoded prediction residual frame can be added to the recovered predicted data to generate the recovered first image frame. As shown in Intra-decoding the digital part of the first encoded data can be accomplished according to any suitable digital coding standard that is employed in intra-encoding the high-information portion of the first image frame at 1040. Intra-decoding the digital part of the first encoded data can include applying entropy decoding, inverse quantization, inverse transform, and prediction to the digital part of the first encoded data. The implementation of the entropy decoding, inverse quantization, inverse transform, and prediction processes are similar to those shown in At 1330, the low-information portion of the first image frame is recovered from the analog part of the first encoded data. In some embodiments, the low-information portion of the first image frame can be recovered by analog-decoding the analog part of the first encoded data. Analog-decoding the analog part of the first encoded data can include the inverse transform process. The inverse transform process can transform the frequency components back to pixel values of the low-information portion of the first image frame. An inverse transform algorithm corresponding to the transform algorithm employed for analog-encoding the low-information portion of the first image frame can be used here. For example, if the DCT is used in analog-encoding the low-information portion of the first image frame, the inverse DCT can be used to obtain pixel values of the low-information portion of the first image frame. In some embodiments, a substitution process can be performed before inverse transform. For example, when the DCT components that do not contribute to the information in the low-information portion of the first image frame were discarded at 1060, the coefficients of the discarded DCT components can be substituted by zero. At 1350, the first image frame is recovered by combing the high-information portion of the first image frame and the low-information portion of the first image frame. Referring again to The inter-decoding process includes applying entropy decoding, inverse quantization and inverse transform, and prediction to the second encoded data. In the entropy decoding process, the second encoded data is converted into decoded quantized transform coefficients. An entropy decoding technique corresponding to the entropy encoding technique employed for inter-encoding the second image frame can be used here. In the inverse quantization process, the decoded quantized transform coefficients are multiplied by the quantization step size (Qstep) to obtain decoded transform coefficients. In the inverse transform process, the decoded transform coefficients are inversely transformed to generate decoded prediction residual data. An inverse transform algorithm corresponding to the transform algorithm employed for inter-encoding the second image frame can be used here. In the prediction process, predicted data can be generated with reference to the recovered frame of the digital part of the first encoded data according to a prediction mode. A prediction mode corresponding to the inter-prediction mode employed for inter-encoding the second image frame may be used. The implementation of the prediction process is similar to the implementation of the inter-prediction process 804 described above. The detailed description thereof is omitted here. In some embodiments, generating the recovered frame of the digital part of the first encoded data including applying entropy decoding, inverse quantization and inverse transform, and prediction to the digital part of the first encoded data. In some embodiments, when the digital part of the first encoded data is generated according to the high-information portion of the first image frame as shown in The decoded prediction residual data can be added to the predicted data to recover the second image frame. At 1150, third encoded data is inter-decoded with reference to the recovered second image frame. The processes of inter-decoding the third encoded data with reference to the recovered second image frame to generate the third image frame are similar to the processes of inter-decoding the second encoded data with reference to the recovered frame of the digital part of the first image frame at 1240. The detailed description thereof is omitted here. In some embodiments, any encoded data after the second encoded data and between the first encoded data and encoded data of a second I-frame, i.e., an I-frame following the first I-frame without another I-frame therebetween, can be inter-decoded with reference to a previously decoded image frame. Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. It is intended that the specification and examples be considered as exemplary only and not to limit the scope of the disclosure, with a true scope and spirit of the invention being indicated by the following claims. An image encoding method comprises hybrid digital-analog (HDA) encoding a first image frame to generate first encoded data including a digital part and an analog part and inter-encoding a second image frame according to the digital part of the first encoded data to generate second encoded data. 1. An image encoding method, comprising:
hybrid digital-analog (HDA) encoding a first image frame to generate first encoded data including a digital part and an analog part; and inter-encoding a second image frame using a reference frame reconstructed from the digital part of the first encoded data to generate second encoded data. 2. The image encoding method according to generating the digital part of the first encoded data from low frequency components of the first image frame; and generating the analog part of the first encoded data from high frequency components of the first image frame. 3. The image encoding method according to employing an intra-prediction on the first image frame to obtain prediction residual data; transforming the prediction residual data into transform coefficients; generating the analog part of the first encoded data according to the transform coefficients corresponding to the high frequency components; quantizing the transform coefficients corresponding to the low frequency components to generate quantized transform coefficients; and entropy encoding the quantized transform coefficients to generate the digital part of the first encoded data. 4. The image encoding method according to determining a threshold to divide the first image frame into the low frequency components and the high frequency components according to at least one of a channel bandwidth, a bit rate, or a resolution of the first image frame. 5. The image encoding method according to generating the digital part of the first encoded data from a high-information portion of the first image frame; and generating the analog part of the first encoded data from a low-information portion of the first image frame. 6. The image encoding method according to generating the digital part of the first encoded data by intra-encoding the high-information portion of the first image frame. 7. The image encoding method according to generating the analog part of the first encoded data by analog-encoding the low-information portion of the first image frame. 8. The image encoding method according to dividing the first image frame into the high-information portion and the low-information portion. 9. The image encoding method according to dividing the first image frame into a plurality of blocks; calculating amounts of information in the plurality of blocks; and assigning the plurality of blocks to the high-information portion of the first image frame or the low-information portion of the first image frame according to the amounts of information of the plurality of blocks. 10. The image encoding method according to inter-encoding a third image frame according to the second encoded data to generate third encoded data. 11. The image encoding method according to reconstructing the second encoded data to obtain a reference frame, and inter-encoding the third image frame according to the reference frame to generate the third encoded data. 12. An encoder, comprising:
a processor; and a memory coupled to the processor and storing instructions that, when executed by the processor, cause the processor to:
hybrid digital-analog (HDA) encode a first image frame to generate first encoded data including a digital part and an analog part; and inter-encode a second image frame using a reference frame reconstructed from the digital part of the first encoded data to generate second encoded data. 13. The encoder according to generate the digital part of the first encoded data from low frequency components of the first image frame; and generate the analog part of the first encoded data from high frequency components of the first image frame. 14. The encoder according to employ an intra-prediction on the first image frame to obtain prediction residual data; transform the prediction residual data into transform coefficients; generate the analog part of the first encoded data according to the transform coefficients corresponding to the high frequency components; quantize the transform coefficients corresponding to the low frequency components to generate quantized transform coefficients; and entropy encode the quantized transform coefficients to generate the digital part of the first encoded data. 15. The encoder according to determine a threshold to divide the first image frame into the low frequency components and the high frequency components according to at least one of a channel bandwidth, a bit rate, or a resolution of the first image frame. 16. The encoder according to generate the digital part of the first encoded data from a high-information portion of the first image frame; and generate the analog part of the first encoded data from a low-information portion of the first image frame. 17. The encoder according to generate the digital part of the first encoded data by intra-encoding the high-information portion of the first image frame. 18. The encoder according to generate the analog part of the first encoded data by analog-encoding the low-information portion of the first image frame. 19. The encoder according to divide the first image frame into the high-information portion and the low-information portion. 20. An unmanned aerial vehicle (UAV) comprising:
a fuselage; a propulsion system coupled to the fuselage and including one or more propellers, one or more motors, and an electronic governor; an image acquiring device coupled to the fuselage and configured to acquire a first image frame and a second image frame; and a processor configured to encode the image by:
hybrid digital-analog (HDA) encoding the first image frame to generate first encoded data including a digital part and an analog part; and inter-encoding the second image frame using a reference frame reconstructed from the digital part of the first encoded data to generate second encoded data.CROSS-REFERENCE TO RELATED APPLICATION
COPYRIGHT NOTICE
TECHNICAL FIELD
BACKGROUND
SUMMARY
BRIEF DESCRIPTION OF THE DRAWINGS
DETAILED DESCRIPTION OF THE EMBODIMENTS










