CA2107727C - Synchronous compression and reconstruction system - Google Patents

Synchronous compression and reconstruction system

Info

Publication number
CA2107727C
CA2107727C CA002107727A CA2107727A CA2107727C CA 2107727 C CA2107727 C CA 2107727C CA 002107727 A CA002107727 A CA 002107727A CA 2107727 A CA2107727 A CA 2107727A CA 2107727 C CA2107727 C CA 2107727C
Authority
CA
Canada
Prior art keywords
video data
reconstruction
video
data
reconstructing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
CA002107727A
Other languages
French (fr)
Other versions
CA2107727A1 (en
Inventor
Hiroaki Ueda
Hidehiro Matsumoto
Yushi Niwa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Personal Computers Ltd
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP4293716A external-priority patent/JPH0822065B2/en
Priority claimed from JP1603593A external-priority patent/JP2500579B2/en
Application filed by NEC Corp filed Critical NEC Corp
Publication of CA2107727A1 publication Critical patent/CA2107727A1/en
Application granted granted Critical
Publication of CA2107727C publication Critical patent/CA2107727C/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/24Systems for the transmission of television signals using pulse code modulation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • H04N21/43072Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/24Systems for the transmission of television signals using pulse code modulation
    • H04N7/52Systems for transmission of a pulse code modulated video signal with one or more other pulse code modulated signals, e.g. an audio signal or a synchronizing signal
    • H04N7/54Systems for transmission of a pulse code modulated video signal with one or more other pulse code modulated signals, e.g. an audio signal or a synchronizing signal the signals being synchronous
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
    • H04N9/8047Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction using transform coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/78Television signal recording using magnetic recording
    • H04N5/781Television signal recording using magnetic recording on disks or drums

Abstract

A synchronous compression system for compressing input video data and audio data with establishing synchronization therebetween, comprises video data compressing means for compressing each of first to nth blocks where n is an integer greater than or equal to two in order of the first block to the nth block, audio data compressing means for compressing input audio data for one frame, interruption control means for controlling to interrupt video data compression when video data compression process is delayed with respect to audio data compression process with reference to frame numbers of those currently compressed in the video data compressing means and the audio data compressing means.

Description

.

SYNCHRONOUS COMPRESSION AND RECONSTRUCTION SYSTEM

BACKGROUND OF THE INVENTION
The present invention relates generally to a synchronouscompressionandreconstructionsystem forvideo and audio data. More specifically, the invention relates to a synchronous compressive recording system for recording video and audio data through compression process with establishing synchronization therebetween, on a recording medium, such as a CD-ROM, hard disk and so forth, and further to a synchronous reconstruction system for synchronously reconstructing video and audio data from the recording medium, onwhich thedata isrecorded throughdata compression process, or through telecommunication circuit.
In the recent years, on-line data retrieval system has been progressively spreading. Through such on-line data retrieval system, video and audio data demanded by users can be retrieved by establishing a database with preliminarily recording video data and/or audio data in a large capacity recording medium, such as CD-ROM, hard disk or so forth and by permitting -the user to access the database through a telecommunication circuit from a terminal.
In such case, the video and audio data are basically analog data. Therefore, it becomes necessary to perform digital conversion for such data for recording on the recording medium. However, when digital conversion is 2107~27 ,~

performed forthese analogdata, amountofdata becomesmuch greater than character data. Therefore, various data compression technologiesare employed forthe video signals and audio signals for recording the large amount of data on the recording medium. Electronic data file systems which can record and retrieve such data have been put into field use.
An example of such system is a multi-media computer system which performs video and audio signal compressing process according to an image coding method in conformance with IS010918-1 (JPEG: Joint Photographic Expert Group). In this compression system, compression of an image is generally performed by dividing one frame into 8 x 8 blocks, and performing discrete cosine transformation (DCT), quantization and entropy coding, respectively per block. Also, reconstruction of the compressed image is performed by performing respective entropy decoding, reverse quantization and reverse DCT for each block of 8 x 8 blocks.
For such system, an apparatus which can perform high speed edition, compression and recording on the recording medium for a large amount of data of video and/or audio and so forth, is required. Also, upon reconstruction, for reconstructing audio data in combination with the correspondingvideodata,matchingofreconstructiontiming (synchronization) becomes necessary. Particularly, due to performance of the apparatus or size of compressed code of 2iO7727 the image, recording and reconstructing speed of the video data can be lower than the recording and reconstructing speed of the audio data so that synchronization cannot be established between the video data and audio data upon reconstruction to give unnatural feeling.
As prior art for establishing synchronization have been disclosed in Japanese Unexamined Patent Publication (Kokai) No. 63-168745, Japanese Unexamined Patent Publication No. 63-266579 and Japanese Unexamined Patent Publication No. 63-313384. These conventional synchronization system will be briefly discussed hereinafter. In the process of data compressive transformation of a large amount of video and audio data, a data flle is established for attribute of data, such as data type, data length, data conversion method, associated data to be synchronized and so forth. These are stored as a data control file. Thereafter, using the data file stored in the data control file, so-called simulated reconstruction in a form assuming reconstruction of data recording in a final form of recording medium is performed. At this time, real time reconstruction of the audio signal, combination of the image and voice, matching qualityof reconstruction timings and so forth are checked by eye and ear, At this time, reconstruction is performed on the basis of the content of data file in a control file, process of simulated reconstruction is modified to freely set the timing or so forth when synchronization between the image and voice is not established. By such simulated reconstruction, operation is checked and adjusted to generate an edited program file.
Insuchvideo/audiosynchronizationsysteminthe conventional video and audio compressive reconstruction system, drawbacks are encountered in that data file mush be generated for matching the timing, and simulated reconstruction on the basis of the content of the data file and checking of operation have to be repeated, to make such work time consuming.
Also, in another conventional video and audio synchronization system, digital codings of the video data and the audio data are performed by mutually independent equipments. Then, the video data and the audio data are correlated according to a given format, transmitted or accumulated, and reproduced byindependent decoders for the video data and the audio data after separating again. As a method for realizing coincidence between the video data and the audio data, a method for providing a reconstruction formatgenerated with predictionofdelaymagnitude between the video data and the audio data in each individual reconstruction system for coupling the video and audio data for simultaneous reconstruction, or a method for providlng an "output delay circuit" for the reconstruction system and 25- adjusting the delay magnitude of a subject dynamic image, is employed.
For example, Japanese Unexamined Patent Publication No. 3-150552, for "Video and Audio Editing System" discloses a synchronization method for performing edition of the audio data recording on a magnetic tape through digital conversion therefor, with respect to a video film, primarily a movie or so forth. In the disclosed method, after performing patching work for the audio data according to an address data included in the magnetic film, reconstruction is performed employing an editing magnetic tape capable of perforation synchronous driving with the video film. Namely, synchronization of the system is realized by hardware, which functions satisfactory with sufficient preliminary adjustment.
Furthermore, an editing means disclosed in the last-identified publication, is provided with means which is descried"having afunctiontomodifyonlylength without causing variation of tone of the voice" as a measure for the case where "words finishes earlier than picture" possibly caused in after recording of "words", so that output image and voice are checked by human eye and ear to make correction with one or more operations.
In the above-mentioned conventional video and audio synchronization system, since amount of data to be generated is not constant depending upon the nature of the subject image, i.e. complicatedness of the picture or magnitude of action and fluctuates per unit for coding so that the period required for decoding reconstruction of the image cannot be determined unltary, synchronization has ,_ been given up initially. Then, measure is taken to predict a delay magnitude as a typlcal value to process all data with a uniform reference, or, as in the editing operation of movie, after checking deviation magnitude between the image 5 and the voice upon reconstruction through actual reconstruction, prediction of the delay magnitude between the image and voice is performed again for coupling in order to obtain coincidence. Therefore, it is not possible to make correction of the delay magnitude between the image and 10 the voice in a method adapted to a condition in real time for precisely maintain synchronization between the image and voice upon reconstruction. This may cause problems of interruption of a sequential reproduced voice, variation of tone and so forth.

SUMMARY OF THE INVENTION
Therefore, it is an object of the present invention to provide a compression system which can perform recording of video and audio data with automatically establishing synchronization of an image and a voice 20 without requiring generation of a data file or so forth.
Another object of the invention is to provide a reconstruction system which can reproduce the video and the audio clata recorded on a recording medium with data compression or transmitted through a telecommunication 25 circuit, with automatically synchronizing the image and the voice without requiring generation of the data file or so forth or cumbersome operations, such as simulated ~ - 7 -Z ~ O ~ 7 2 7 ~
reconstruction or so forth.
A further object of the present invention is to provide a synchronous reconstruction system which permits correction of a deviation magnitude to be caused during decoding reconstruction of digital coded video and audio data in a method adapted to the condition in real time.
According to a first aspect of the present invention, a synchronous compression system for compressing input video data and audio data with establishing synchrcnization there-between, comprising: video data compressing means forcompressing each of first to nth blocks of video data where n is an integer greater than or equal to two in order of the first block to the nth block; audio data compressing means for compressing input audio data for one frame; interruption control means for providing control by interrupting video data compression when compression of video data is delayed with respect tc compression of audio data with reference to frame numbers of those frames currently compressed in said video data compressing means and said audio data compressing means.
According to a second aspect of the invention, a synchronous reconstruction system for synchronously reconstructing input video data and audio data respectively compressed per frame, comprises: video data reconstructing means for scanning first to nth blocks of one frame input video data in order, A

.

which each one frame of video data is divided into first to nth blocks;
audio data reconstructing means for reconstructing input audio data for one frame;
interruption control means for controlling to interrupt video data reconstruction when video data reconstruction process is delayed with respect to audio data reconstruction processwith reference to frame numbers of those currently compressed in the video data reconstructing means and the audio data reconstructing means.
According to a third aspect of the invention, a synchronous reconstruction system for synchronously decoding and reconstructing digitally coded video data and audio data, comprises:
storagemeansforpreliminarilystoringdigitally coded video and audio data;
video data reconstructing means for decoding and reconstructingdigitallycodedvideodataperpredetermined data amount;
audio data reconstructing means for decoding and reconstructing audio data;
transfer control means for reading out digitally coded audiodata from the storagemeans and transferring the read out digitally coded audio data to respective of the video data reconstructing means and the audio data reconstructing means: and ' -~ 2107727 control means responsive to completion of decoding and reconstructing process of both of the video data reconstructingmeans and theaudio data reconstructing means for controlling modification of initiation timing of the video data reconstructing means and image display period after video reconstruction process.

BRIEF DESCRIPTION OF THE DRAWINGS
The present invention will be understood more fully from the detailed description given herebelow and from the accompanying drawings of the preferred embodiment of the invention, which, however, should not be taken to be limitative to the invention but are for explanation and understanding only.
In the drawings:
Fig. 1 is a block diagram of a video and an audio synchronouscompressionsystemapplicableforanembodiment of the present invention;
Fig. 2 is a block diagram of a video and an audio synchronous reconstruction system applicable for an embodiment of the invention;
Fig. 3 is a flowchart of an operation of the system of Fig. 1;
Fig. 4 is a flowchart of an operation of the system of Fig. 2;
Fig. 5 is an illustration showing an example of the manner for dividing one frame of image into a plurality of blocks;

~10~727 Fig. 6(A) i s a timing chart in compression of one frame image per block;
Fig. 6(B) i s an illustratlon of a data format after compression;
Fig. 7 is a flowchart showing the compression process of Fig. 6;
Fig. 8 i s a flowchart showing the reconstruction process of Fig. 6;
Fig. 9(A) is an illustration showing an example of the case where one block of the image is further divided into a plurality of frequency components;
Fig. 9(B) i s an illustration showing an example of a transforming equation for compressing and reconstructing process of the image per the frequency component;
1 5 Fig. 10(A) is a timing chart in the case where compression process is performed per the frequency component of the one frame of image;
Fig. 10(B) is an illustration of a data format after compression;
Fig. 11 is a flowchart showing the compression process of Fig. 10;
Fig. 12 i s a flowchart showing the reconstruction process of Fig. 10;
Fig. 13 i s an illustration showing an example, in which the one block of image is further divided into a plurality of chrominance components;
Fig. 14(A) is a timing chart for compressing the one frame of image per chrominance component:
Fig. 14(B) is an illustration of a data format after compression:
Fig. 15 is a flowchart showing compression process of Fig. 14;
Fig. 16 i s a flowchart showing reconstruction process of Fig. 14;
Fig. 17 is a block diagramof another embodiment of the present invention;
1 0 Fig. 18 i s a flowchart showing an operation of the blocks of Fig. 17 Figs. 19 ~ 25 are illustrations showing a relationship between an image frame and a voice: and Fig. 26 i s a block diagram of a further embodiment of the present invention.
DESCRIPTION O F THE PREFERRED EMBODIMENT
The preferred embodiments of the present invention will be discussed hereinafter in detail with reference to the drawings.

Fig. 1 is a block diagram of a system applicable for the preferred embodiment of a synchronous compression system according to the present invention, which illustrates a video and an audio compression system in an image coding method of a multi-media computer in conformance with IS010918-1 (JPEG).
In the drawings, CPU 1 executes programs stored in ROM 2 for controlling the overall system. RAM 3 is a memory for temporarily storing data during compression process. A key board 4 is adapted to make designation of video and audio data compression process. A hard disk 5 records videocodeand audiocodecodedwith compression and stored in the RAM 3. A transmitter portion 6 transmits the coded and compressed video code and audio code to a telecommunication circuit.
An input video signal is subject to A/D
conversion by an A/D converter 8 and separated into a luminance signal and a chrominance signal by a Y/C
separator 7. Thereafter, the input video signal is subject to a discrete cosine transformation by a DCT transformer 9, quatization by a quantizer 10 and high efficiency compression and coding by a codert1, and then stored in the RAM 3. An input audio signal is subject to compression and coding by an audio coder 12 and stored in the RAM 3.
Fig. 2is a system block diagram applicable for the preferredembodimentofasynchronousreconstructionsystem of the present invention and showing a video and audio reconstructing apparatus employing the image coding method of a multi-media computer in conformance with IS010918-1 (JPEG).
In Fig. 2, CPU 20 executes programs stored in ROM
21 for controlling the overall system. RAM 22 temporarily stores data during reconstruction process. A key board 23 performs process of a video and audio reconstructing apparatus. A hard disk 24 records compressed and coded 21077~7 video code and audio code, which codes can be read in the RAM 22. A receiver portion 23 receives the compressed and coded video code and audio code.
The compressed video code read in the RAM 22 is coded by a decoder 30 and subject to a reverse quantization by a reverse quantizer 29. Thereafter, reverse discrete cosine transformation by a reverse DCT transformer 28, composition of the luminance signal and the chrominance signal by a Y/C composer, and D/A conversion by a D/A
converter 27 are performed for outputting a video signal.
The compressed audio signal read in the RAM 22 is decoded by an audio decoder 31 and output as an audio signal.
Fig. 3 is a flowchart for controlling compression process for the video and audio data. As shown in the flowchart of Fig. 3, once the compression process is initiated, ON/OFF of a transmission flag, number of frames (or fields) to be recorded, file and so forth are input through the key board 4 (step 40). Then, the input video signal is converted into a digital data by the A/D
converter 8 (step 41), separated into the luminance signal and the chrominance signal by the Y/C separation circuit 7 (step 42), subject to video compression (step 43) and stored in the RAM 3 as compressed code of the video data.
Next, theinput audio signal iscoded by the audio coder 12 (step 45), and stored in the RAM 3 as the audio compressed code (step 46). Check is performed whether the transmission flag is ON (step 47). Otherwise, the process is advanced to a step 49. When the transmission flag is ON, the video and audio codes in the RAM 3 are transmitted to the telecommunication circuit through the transmitter portion 6 (step 48). Subsequently, the video and audio codes in the RAM 3 are recorded on the hard disk 5 (step 49).
Then,judgementismadewhetherthecurrent frametor field) is the final frame (or field) (step 50). If not, the process returns to the step 41 and otherwise, i.e. when the current frame is the final frame, process goes end.
Fig. 4 is a flowchart of a routine for controlling video and audio reconstruction process. As shown in the flowchart of Fig. 4, once a reconstructing apparatus initiates operation, ON/OFF of a reception flag, number of frames (or fields) to be reproduced and file name are input through the key board 23 (step 60). Next, check is performed whether the reception flag is ON (step 61). If so, the compressed video and audio codes are received from the telecommunication circuit by the receiver portion 25 and stored in the RAM 22 (step 62). The compressed video and audio codes stored in the RAM 22 is recorded on the hard disk 24 (step 63).
If the reception flag is not ON, the compressed video and audio codes are read out from the hard disk 24 and stored in the RAM 22 (step 64). Then, the process is advanced to a step 65. Then, the compressed audio code stored in the RAM 22 iS decoded by the audio decoder 31 for outputting as the audio signal(step 65). Subsequently, the 210~727 compressed video code is reproduced (step 66). The luminance signal and the chrominance signal are then composed by the Y/C composer 27 (step 67), and converted by the D/A converter 27 fir outputting as the video signal (step 68). Next,judgement isperformed whether the current frame (or field) is the final frame (or field), (step 69.
If not, the process returns to the step 61, and otherwise, i.e. when the current frame is the final frame, the process goes end.
10The foregoing is the brief of the construction and operation of the compression and reconstruction system of the preferred embodiment of the invention. The further detail of the preferred embodiments will be given hereinafter, which discussion will be given separately.
15(1) Interrupting Operation per Unit of Video Compression and Reconstruction Process Figs. 5 ~ 8 are illustrations for an example of process forvideoand audio synchronizationby interrupting a video compression and reconstruction process per unit when the video compression and reconstruction are delayed relative to an audio compression and reconstruction process.
Fig. 5 is an illustrationto be used for discussion with respect to the block as a processing unit while one frame of image is compressed. Although one frame is discussed as 48 x 48 dots in Fig.5, it is practically 600 x 400 dots. Here, 48 x 48 will be used for simplification.

~.--The image of 48 x 48 dots are separated into thirty-six blocks of 0th to 35th (each block will consist of 8 x 8 dots). The compression process is performed per the unit of blocks.
5In this case, the order of compression process is zigzag scanning to can the blocks according to the block number in a order of smaller to greater, At this time, as shown in Fig. 6(A), the audio data and the video data in the same frame are processed in parallel. Here, assuming that 10while the 30th block of the 2nd frame is in process, and the audio compression process is initiated for the 3rd frame, the 31st block of the 2nd frame of the video data is not processed for restarting the process from the 0th block in the 3rd frame.
15As shown in Fig. 6(B), in the audio and the video codes thus prepared, the video codes of the 0th to 35th blocks (MO ~ M35) are stored for the first frame, in which interruption has not been caused, and the video code of the 0th - 30th blocks are stored for the second frame. in which 20the interruption is caused.
It should be appreciated, in the drawing, A
denotes the compressed audio signal, SOI is a code representative of the leading end of the frame, and EOI is a code representative of the trailing the end of the frame.
25Fig. 7 is a flowchart showing the process for interrupting video compression per block. As shown in the flowchart of Fig. 7, once video compression is initiated, '- 2107727 judgement is made whether the current frame is the first frame (step 80). If so, the process is jumped to a step 83, and otherwise, the audio frame number and the video frame number currently processed are compared each other (step 81) to make judgement whether the video processing is delayed to the audio recording(step 82). If delayed, after completion of compression for the currently processed block, then video signal processing for the current frame is terminated. Otherwise, the frequency component is derived from the block by the DCT transformer (step 83).
Then quantization by the quantizer 10 (step 84), high efficiency compressive coding by the coder 11 (step 85) and storing the compressed video code is stored in the RAM 3 (step 86), are performed. Finally, check is performed whether the current block is the final block or not.
Upon reconstruction, as shown in Fig. 8, once video reconstruction is initiated, check is initially performed whether the current frame is the first frame (step 90). It so, the process is advanced to a step 93.
Otherwise, comparison is made between the audio frame number and the video frame number (step 91) to make judgement whether the video processing is delayed from the audio processing or not (step 92). If delayed, after completion of compression for the currently processed block, then video signal processing for the current frame isterminated. Otherwise, the compressed video code stored in the RAM 22 is read out (step 93) and decoded by the 21~7727 " _ .

decoder 30 (step 94). Then reverse quantization by the reverse quantizer 29 (step 95), and the value of the image is derived by the reverse DCT transformer 28 (step 96).
Finally, the check is performed whether the current block is the final block or not (step 97). Otherwise, the process returned to the step 90. If the current block is the final block, reconstruction of the video data is terminated.
(2) Interruption Operation per Frequency Component of Video Compression and Reconstruction Process Figs. 9 ~ 12 is an illustration showing an example of process for establishing synchronization between the audio and video by interrupting video compression and reconstruction process per every frequency components in the unit of the block when the video compression and reconstruction process is delayed to from the audio compression and reconstruction process, Fig. 9(A) shows that with respect to each block (one block) of the 0th to 35th shown in Fig. 5, combination of horizontal frequency 8 x vertical frequency 8 (64 components) are further divided into four components of 4 (O - 3).
As an order for compression and reconstruction process, process is performed in order to the 0th ~ 63th frequency components of 0th block, 0th - 63th frequency components of the 1st block, 0th ~ 63th frequency components of the 2nd block and 0th to 63th frequency components of the 3rd block.
Fig. 9(B) showsatransformingequationsuponvideo - 2ln7727 compression and reconstruction. In the figure, Svu is a sample value after compression, syxis a sample value before compression or after reconstruction, ~ is a horizontal spatial frequency, and v is a vertical spatial frequency.

As shown in Fig. 1 O(A), the audio data and video data in the same frame are processed in parallel. Here, assuming that while the 30th frequency component of the 3rd block component of the 2nd frame is in process, and the audio compression process is initiated for the 3rd frame, the 31st frequency component of the 3rd block of the 2nd frame is not processed for restarting the process from the 0th frequency component in the 0th block of the 3rd frame.
The audio and video codes thus generated are stored. In the shown case, as shown in Fig. 10(B), the codes (HOOO ~ H063) of the 0th ~ 63th frequency components of the 0th block, the codes (H100 ~ H163) of the 0th ~ 63th frequency components of the 1st block, the codes (H200 ~
H263) of the 0th ~ 63th frequency components of the 2nd block, the codes (H300 ~ H363) of the 0th ~ 63th frequency components of the 3rd block, of the 1st frame, in which interruption is not caused, are stored. For the 2nd frame, in which the interruption is caused, the codes (HOOO ~ H263) of the 0th ~ 63th frequency components of the 0th ~ 2nd blocks, and the codes (H300 ~ H330) of the 0th ~ 30th frequency components of the 3rd block are only stored.
Fig. 11 shows a flowchart for interrupting process for the video compression process per a unit of the frequency component. As shown in the flowchart of Fig. 11, once video compression is initiated, judgement is made whether the current frame is the first frame (step 100). If so, the process is jumped to a step 103, and otherwise, the audio frame number and the video frame number currently processed are compared each other (step 101) to make judgement whether the video processing is delayed to the audio recording(step102). Ifdelayed, after completionof compression for the currently processed block of the frequency component, then video signal processing for the current frame is terminated. Otherwise, the frequency component is derived from the block by the DCT transformer (step 103). Then quantization by the quantizer 10 (step 104), high efficiency compressive coding by the coder 11 (step 105) and storing the compressed video code is stored in the RAM 3 (step 106), are performed.
Next, check is performed whether the processed frequency component is the final frequency component or not (step107). If not, the process returns to the step100, and otherwise, i.e. the processed frequency component is the final frequency component, the frequency component to be processed is modified (step 108). Then, judgement is made whether the current block is the final block or not (step 109). If not, process is returned to the step 100, and otherwise, i.e. when the processed frequency component is the final frequency component, then video compression is terminated.

~107727 Fig. 12 is a flowchart for video reconstruction process. Upon reconstruction, as shown in Fig. 12, once video reconstruction is initiated, check is initially performed whether the current frame is the first frame ( step 120) . It so, the process is advanced to a step 123.
Otherwise, comparison is made between the audio frame number and the video frame number currently processed (step 121) to make judgement whether the video processing is delayed from the audio processing or not (step 122 ~ . If delayed, after completion of compression for the currently processed block, then video signal processing for the current frame is terminated. Otherwise, the compressed video code stored in the RAM 22 is read out (step 123) and decoded by the decoder 30 (step 124) . Then reverse quantization by the reverse quantizer 10 tstep 125), and the value of the image is derived by the reverse DCT
transformer 28 (step 126) .
Next, the check is performed whether the processed frequency component is the final frequency component or not (step 127). If not, the process returns to the step 120, and otherwise, i.e. the processed frequency component is the final frequency component, the frequency component to be processed is modified (step 128). Then, judgement is made whether the current block is the final 25 block or not (step 129). If not, process is returned to the step 120, and otherwise, i.e. when the current block is the final block, then video reconstruction is terminated.

~ln7727 (3) InterruptionOperationperChrominanceComponent of Video Compression and Reconstruction Process Figs. 13 ~ 15 is an illustration showing an example of process for establishing synchronization between the audio and video by interrupting video compression and reconstruction process per chrominance components in the unit of the block when the video compression and reconstruction process is delayed to from the audio compression and reconstruction process, Fig. 13 shows an example that with respect to each block (one block) of in Fig. 5, is further divided into three components, i.e. Y
component, Cbcomponent andCrcomponent. Then, theprocess is performed in an order of the 0th ~ 35 blocks of 0th chrominance component, 0th - 35th blocks of the 1st chrominance component and 0th to 35th blocks of the 2nd chrominance component.
As shown in Fig. 14(A), the audio data and video data in the same frame are processed in parallel. Here, assuming that while the 30th block of the 2nd chrominance component of the 2nd frame is in process, and the audio compression processis initiated for the3rd frame, the31st block of the 2nd chrominance component of the 2nd frame is not processed for restarting the process from the 0th block in the 0th chrominance component of the 3rd frame.
The audio and video codes thus generated are stored. In the shown case, as shown in Fig. 14(B), the codes (COOO ~ C035) of the 0th ~ 35th blocks of the 0th .. .

chrominance component, the codes (ClOO ~ C135) of the 0th ~
35th blocks of the 1st chrominance component, the codes (C200 - C235) of the 0th ~ 35th blocks of the 2nd chrominance component of the 1st frame, in which interruption is not caused, are stored. For the 2nd frame, in which the interruption is caused, the codes (COOO ~ Cl35) of the 0th ~ 35th blocks of the 0th ~ 1st frequency components, and the codes (H200 ~ H230) of the 0th ~ 30th blocks of the 2nd chrominance component are only stored.
o Fig. 15 shows a flowchart for interrupting process for the video compression process per a unit of the chrominance component. As shown in the flowchart of Fig. 11, once video compression is initiated, judgement is made whether the current frame is the first frame (step 140). If so, the process is jumped to a step 143, and otherwise, the audio frame number and the video frame number currently processed are compared each other (step 141) to make judgement whether the video processing is delayéd to the audio recording (step 142). If delayed, after completion of compression for the currently processed block of the chrominance component, then video signal processing for the current frame is terminated. Otherwise, the chrominance component is derived from the block by the DCT transformer (step 143). Then quantization by the quantizer10 (step144), highefficiencycompressivecoding by the coder 11 (step 145) and storing the compressed video code i~ stored in the RAM 3 (step 146), are performed.

Next, check is performed whether the current block is the final block or not (step 147). If not, the process returns to the step 140, and otherwise, i.e. the current block is the final block, the block is returned to the leading block and the chrominance component to be processed is modified (step 148). Then, judgement is made whether the processed chrominance componentis the final chrominancecomponent or not(step149). If not, process ls returned to the step 140, and otherwise, i.e. when the processed chrominance component is the final chrominance component, then video compression is terminated.
Fig. 16 is a flowchart for video reconstruction process. Upon reconstruction, as shown in Fig. 16, once video reconstruction is initiated, check is initially performed whether the current frame is the first frame (step 160). It so, the process is advanced to a step 163.
Otherwise, comparison is made between the audio frame number and the video frame number currently processed (step 161 ) to make judgement whether the video processing is delayed from the audio processing or not (step 162). If delayed, after completion of compression for the currently processed block, then video signal processing for the current frame is terminated. Otherwise, the compressed video code stored in the RAM 22 is read out (step 163) and decoded by the decoder 30 (step 164). Then reverse quantization by the reverse quantizer10(step 165), and the value of the image is derived by the reverse DCT

'~ 21Q7727 transformer 28 (step 166).
Next, the check is performed whether the current block is the final block or not (step 167). If not, the process returns to the step 160, and otherwise, i.e. the current block is the final block, the block is returned to the leading block and the chrominance component to be processed is modified (step 168). Then, judgement is made whether the processed chrominance component is the final chrominance component or not(step169). If not, process is returned to the step 160, and otherwise, i.e. when the processed chrominance component is the final chrominance component, then video compression is terminated.
It should be noted that the input video data and the input audio data (data read out from the recording medium or data received through the telecommunication circuit) in the shown embodiment of the video and audio reconstruction process are employed as data not effected synchronization process upon video and audio compression process and synchronization process is performed upon reconstruction therefor.
Next, another embodiment of a synchronous reconstruction system according to the present invention will be discussed. In this embodiment, a reconstruction time axis for the audio data is assumed as conceptual operation time of the overall system. Then, a deviation of displaying timings of an actual video frame displayed at audio reconstruction processing period for a unit data '~ 2107727 amount and an ideal frame estimated to be displayed at the same timing determined by calculation, is measured to establishsynchronizationofthesequential image and voice by modifying the display timing of the actual frame.
It should be appreciated that the video frame represents one still image in handling dynamic image.
Assuming that the video data is converted into numerical data through coding process, it will be referred to as coded image frame. Also, this unit is simply referred to as frame. The actual frame means an actually displayed frame at an arbitrary designated timing of the dynamic image reproduced with depending upon the operation speed of the video decoding reconstruction apparatus and nature of the image. Ideal frame means the video frame to be displayed at arbitrary designated timing when the reconstruction processing speed of the audio data is constant and does not vary within the unit period and the period required for composite reconstruction of the image is sufficiently short.
The construction of the shown embodiment is illustrated in Fig. 17. The shown embodiment of the synchronous reconstruction system includes a coded data storagedevice171 forstoringpreliminarilycodedvideo and audio data, a synchronization control device 173 receiving the video and audio data from the storage device 171, controlling the mutual output timing of the image and voice with reference to a timemeasuringdevice172 and outputting ~1~7727 the video data in a data amount of one frame and the audio data in a data amount corresponding to the one frame of coded video data. The system further includes a coded video data decoding and reconstructing apparatus 174 for decoding and reconstructing every one frame of video data output from the synchronization control device 173, a coded audio data decoding and reconstructing apparatus 175 for decoding and reconstructing the audio data of a data amount arbitrary given by the synchronization control device 173, and an output device 176 for receiving the reconstruction outputs of the coded video data decoding and reconstructing apparatus 174 and the coded audio data decoding and reconstructing apparatus 175 to convert then into visible and audible data.
Furthermore, the synchronization control device 173 is constructed as a software which includes a trans~er control means 177 for reading out video and audio data from the storage device 171 and transferring then to the coded videodatadecodingandreconstructingapparatus 174 and the coded audio data decoding and reconstructing apparatus 175, a decoding and reconstruction apparatus initiation control means 178 for controlling the coded video data decoding and reconstructing apparatus 174, thecoded audio data decoding and reconstructing apparatus 175 and the transfer control means 177 with reference to the time measuring device 172, a setting means for setting an interruption time interval I of the time measuring device 172. Similarly, the time ,, measuring device172 is constructed by a software including an interruption means 180 for effecting interruption and a timer 181 for measuring an arbitrary period.
Here,thepresentinventionperformsdecodingand reconstruction of data by control within the apparatus in response to an operation initiation command provided in a condition where the coded data to be decoded and reproduced is stored in the buffer memory, without providing constraint for the digital coding method of the video and audio signals handled by the coded video data decoding and reconstructing apparatus 174 and the coded audio data decoding and reconstructing apparatus and by providing the buffer memory having a sufficient capacity for desired functions. This process is adapted to perform reconstruction forone frameofthevideo dataand arbitrary determined data amount of the audio data at one time.
In the coded data storage device 171, the recording format of the video and audio data and the coupling method therefor are not specified to enable reading out of data at sufficiently high speed. For instance, time code, format, in which the audio data are arranged at upstream and downstream of the video data may be employed as long as consistency can be judged.
Next, the operation of this embodiment will be discussed. Fig.18 is a flowchart showing the operation of the shown embodiment.
Once operation is initiated, the decoding and - ~iO77~7 reconstructing apparatus initiation control means 178 in the synchronization control device 173 inputs coded audio data read out from the coded data storage device 171 to the data buffer in the coded audio data decoding and reconstructing apparatus 175 for a arbitrary determined unit periodt, at a step S1. At this time, reconstruction of the video data is preliminarily corresponded so that n frames can be reproduced within the unit periodt. Next, at a step S2, the decoding and reconstructing apparatus initiation control means 178 commands starting of generation of voice and starting of displaying of a sequential image for t seconds as the unit period. At a step S3, one frame of coded video data is input to the coded video data reconstructing apparatus 174 via the transfer control means 177 to initiate the decoding and reconstructing function. The input coded data is decoded by internal control of each decoding and reconstructing apparatus and becomes visible and audible by providing to the output device 176 as converted analog voice and audio signals. The operation up to here, displaying of the image and generation of voice for the first frame is initiated.
Next, decoding and reconstructing apparatus initiation control means 178 inputs the second frame coded video data to the coded video data decoding and reconstructing apparatus 174 to trigger decoding and reconstructing function. The coded video data decoding and reconstructing apparatus 174 instantly initiate ~ 2~7727 reconstructing operation and output device 176 displays the second frame in synchronism with its own operational clock.
By repeating this operation for n times, n frames of dynamic image can be reproduced for displaying. Also, by repeating 5 n+l times, the dynamic image of over n frames can be reproduced.
Here, the video data to be used for reconstruction is for n frames, Assuming that there is not fluctuation in the reconstructing function of the video and lO audio data and reconstruction is performed regularly, reconstruction of nth frame of the video data should be accurately completed after t seconds. Then, that timing is the initiation timing of reconstruction of the n+ l frame, as shown in Fig. 19. Here, the n+l frame of the video data 15 will be referred to as "measurement object video frame".
However, it is unlikely to simultaneously complete reconstruction of the video data and the audio data without causing any error. Normally, either of the video and audio data reconstruction is completed at earlier 20 timing than the other. Therefore, there is certain magnitude of difference between the termination timing of reconstruction of the video data and the audio data. In order to measure a timing difference between the audio reconstruction completion timing and the display timing of 25 the video data of the n+l frame, at first, at a step S4, check is performed whether the period t is expired and the reconstruction of the audio da-ta is completed or not. If - 21077~7 not completed, at every time of decoding and reconstructing operation of each frame, check is performed whether the image is the measurement object frame at a step S5. If not, the process returns to the step S3.
When judgement is made that the unit period t is elapsed and reconstruction of the audio data is completed at the step S4, the timer181 is set by the setting means 179 to initiate time measurement, at a step S6. Subsequently, at a step S7, the coded video data is input to the coded video data decoding and reconstructing apparatus 174 to initiate decoding and reconstructing function. At a step S8, at every occurrence of reconstruction of respective frames, check is performed whether the image is the measurement object video frame or not. This is repeated until the measurement object video frame is detected. When the measurement object video frame is detected, the measurement of the time is terminated at a step S9.
On the other hand, despite of the fact that the reconstruction of the audio data at the step S5 is not yet completed, thejudgement is made that the video frame is the ~ measurement object video frame, the timing is set by the setting means 179 at a step S10 to initiate measurement of the time. Subsequently, at a step S11, when judgement is made that the unit periodtis elapsed and reconstruction of the audio signalis completed, themeasurement ofthe period is terminated similarly to the step S8.
It should be noted that the measurement of the elapsed time by the time measuring device 172 employs the interruption means 80 connected to the timer 181. At every time interval i, interruption for the synchronization controlsystem 173 is effected. Thesynchronizationcontrol system 173 counts the number of occurrence of the interruption forrealizingmeasurement ofthe elapsed time.
However, the method for measuring the period is not specified to the shown method. Once measurement of the elapsed time is completed at the step S9, then the delay amount of the video and audio data is set with the elapsed time to perform the following process according to the magnitude of the delay d, at the step S12.
As set forth above, ideally, when the displayed frame after t seconds is the n+lth video frame or the delay magnitude d is within the period required for displaying (within a vertical fly-back period, i.e. d - O, of the output device 176), the coincidence of video and audio data is considered established. Then, synchronization process is terminated at a step S13. At a step S14, the display period of the n+lth frame is expanded or shortened to effect synchronization correcting operation, then the synchronization process is terminated at the step S13.
In detail, when the audio data is delayed, the display timing of the n+2th frame is delayed to expand the display period of the n+lth frame to make the end coincident with the end of the audio reconstruction. On the other hand, when the video data is delayed, the display timing of ~ 2107727 the n+2th frame is advanced to shorten the display period of the n+lth frame to make the end coincident with the end of the audio reconstruction.
At a step S12, when the delay magnitude d is greater than one frame the display period(d > t/n seconds), correction of synchronization is performed by thinning one or more frames or repeating display of the same frame at a step S15, and then synchronization process is terminated at the step S13. In detail, when the audio reconstruction is delayed, displaying of the n+2th frame is waited for expanding the display period of the n+lth frame until the timing where synchronization with the audio reconstruction can be established. On the other hand, when the video reconstruction is delayed, n+2th frame is neglected from displaying to effect adjustment beyond the one frame display period (t/n seconds). Namely, in this case, the order of display is [... n-l, n, n+l, n+3 ...]. If neglecting of displaying of n+2th frame is not sufficient for correction, n+3, 4, 5th frames may be neglected from displaying.
Next, further discussion will be given with concrete numerical values.
Here, in order to realize the video and audio synchronization system on a general purpose personal computer(hereafter referredto as PC), the codedvideo data decoding and reconstructing apparatus 174 and the coded audio data decoding and reconstructing apparatus 175 are realized on an extended board corresponding to the PC.
Also, the output device 176 (one which requires approximately 33 msec. for displaying one frame of video data) is provided. On the other hand, in order to realize 5 the coded data storage device 171 and the synchronization control device 173 with programs, RAM is provided. However, as set forth above, the buffer memories necessary for respective decoding and reconstructing apparatus 174 and 175 are included in respective of the decoding and l0 reconstructing apparatus. Also, the speed of data transfer is sufficiently high.
The system is termed to reproduce a dynamic movie (image and voice) longer than or equal to two seconds. In addition, for convenience of discussion,the coded audio 15 data is assumed as t4 bit, stereo 8.27 Khz of ADPCM
(adaptive differential type pulse code modulation system)], and one second (8270 bite) of the audio data is transmitted as one unit. On the other hand, when reconstruction of one second of audio data is completed, second unit of audio data 20 is continuously transferred to continue reconstruction so that the voice will not be interrupted even in reconstruction of the dynamic movie of longer than or equal to two seconds.
As operation, in conjunction with initiation of 25 decoding and reconstruction of the audio data, the data transfer of the first frame of video data, decoding and reconstruction thereof is initiated. By repeating decoding " 2ln7727 -and reconstructing operation of the video data, decoding and reconstructing frames are counted. When the number of frame reaches the measurement object frame number(the case where the video reconstruction speed is high and audio reconstruction speed is low, will be referred to as pattern A), oraudio reconstructioniscompleted(the casewhere the video reconstruction speed is low and audio reconstruction speed is high, will be referred to as pattern B), the timer 181 (since it is desirable to use convenience value, such as 10 msec as interruption interval of the timer 181, 10 msec.
is used in this example) starts operation.
The timer181 isstopped when the one second audio reconstruction is completed in the pattern A, or the measurement object frame is reached in the pattern B. The number of timer interruption is then calculated and delay magnitude is determined. Here, assuming that the predetermined number of the video frame to be displayed within one second is 15 (hereafter, since t/n 0.66 msec.the following discussion will be given assuming 0.66 msec.), discussion will be given by classifying with the delay magnitude d.
When the delay magnitude d is:
(1) video reconstruction is later than audio reconstruction for 10 msec. (Fig~2o);
(2) video reconstruction is later than audio reconstruction for 50 msec. (Fig.21) (3) video reconstruction is later than audio reconstruction for 70 msec. (Fig.22) (4) audio reconstruction is later than video reconstruction for lO msec. (Fig.23) (5) audio reconstruction is later than video reconstruction for 50 msec. (Fi9.24) (6) audio reconstruction is later than video reconstruction for 70 msec. (Fi9.25).
For respectiveofthecases, mannerofcorrection of synchronization will be discussed hereinafter. It should be noted that Figs. 21, 22, 24 and 25 are illustrated as separated drawings for making the conditions clear, in which(A)illustratesdelaycondition ofthe video and audio reconstruction, and(B)illustratesmannerofcorrection of synchronization.
(l) When Video Reconstruction is later than Audio Reconstruction for lO msec. (Fi9.20) This is the condition where 'd' - 0 is judged at the step S12 of Fig. 18. Therefore, correction of the synchronization is not performed.
(2) When Video Reconstruction is later than Audio Reconstruction for 50 msec. (Fi9.21) This is the case where 0 < 'd' < t/n is judged at the step S12 of Fig. 18, and thus the video reconstruction is delayed. In this case, the display period of 16th frame (n+l) is shortened. In the shown case, the delay magnitude d is greater than display period(33 msec.) for one frame of the output device and smaller than the display period (66 msce) of one frame. Therefore, the display period of the n+lth frame is shorted for 33 msec. Namely, correction of synchronization can be performed by advancing the display timing of the n+2th frame for 33 msec.
(3) When Video Reconstruction is later than Audio Reconstruction for 70 msec. ( Fig. 22 ) This is the condition where t/n < ,d, isjudged at the step S12 of Fig. 18 and the video reconstruction is delayed. In this case, the 17th frame (n+2) is neglected from displaying. IN the shown case, the delay magnitude d is greater than the display period (66 msec.) of one frame.
Therefore, by skipping the 17th frame, the data of the 18th frame (n+3) is transferred to the video data decoding and reconstructing apparatus 174. By this operation, the actually reproduced frames are [1, ..... 15. 16, 18 ... ]
for correction of synchronization.
(4) When Audio Reconstruction is later than Video Reconstruction for 10 msec. ( Fig. 23) This is the condition where 'd' 0 is judged at the step S12 of Fig. 18. Therefore, correction of the synchronization is not performed.

(5) When Audio Reconstruction is later than Video Reconstruction for 50 msec. ( Fig. 24 ) This is the condition where 0 < Id' ~ t/n is judged at the step S1 2 of Fig. 18 and the audio reconstruction is delayed. The display period (66 msec.) 2ln7. 727 of the 16th frame is extended for delaying the decoding and reconstructing timing of the 17th frame (n~2) to make correction of synchronization. This process is differentiated from that of (2) in that, in the former case the video reconstruction is delayed relative to the audio reconstruction, but the audio reconstruction is delayed with respective to the video reconstruction in the shown case.
(6) When Audio Reconstruction is later than Video Reconstruction for 70 msec. ( Fig. 25 ) This is the case where t/n < 'd' is judged at the step S12 of Fig. 18. Similarly to (5), the display period of the 16th frame is expanded. Though the practical process is the same as (5), the display is maintained for longer period. Here, since the display period for one frame is 66 msec., the correction for 70 msec. is a correction for approximately one frame.
As set forth above, depending upon the measured delay amount, the operation for correction of synchronization is performed with respect to each of the foregoing six cases. In the shown example, since reconstruction of the movie (image and voice) for a period longer than or equal to 2sec, number of reproduced frame is subsequently counted to perform similar correction for synchronization is performed at every arbitrary determined unit period t (1 second in the shown case).
Fig. 26 is a block diagram of a further embodiment of the synchronous reconstruction system according to the present invention. Like reference numerals represent like elements to those in Fig.17. In this embodiment, a video and audio decoding and reconstructing system 182 including the coded video data decoding and reconstructing means 174 and the coded audio data decoding and reconstructing means 175 are realized by a software. Other portions are the same as those in Fig.17. Therefore, furtherdiscussion is neglected for simplification of the disclosure.
Here, the method and system for coding the video signal and the audio signal to be handled by the coded video data decoding and reconstructing means 174 and the coded audio data decoding and reconstructing means 175 may be any method or system which can be processed by the software. In order to satisfy such functions, the buffer memory should have sufficient storage capacity. For instance, when the coded data to be decoded and reproduced is stored in the buffer memory realized by the hard disk, decoding and reconstructing operation is performed in response to operation initiation command. This process is to perform reconstruction fortheone frameofvideo data and arbitrary determined data amount of audio data. at one time.
On the other hand, concerning the video data and the audio data stored in the coded data storage device 171, the recording format of the video and audio data and the coupling method therefor are not specified to enable reading out of data at sufficiently high speed. For instance, time code, format, in which the audio data are arranged at upstream and downstream of the video data may be employed as long as consistency can be judged.
As set forth above, according to the present invention, by interrupting video data compression process adapting to the speed of the audio data compression process during video and audio data compression, higher speed in video data compression process can be achieved with automatically establishing synchronization with the audio data compression process. Also in reconstruction, by interrupting video data reconstructing process adapting to the reconstruction speed of the audio data, higher speed is achieved in video data reconstruction process with automatic synchronization with the audio data reconstruction.
As a result, since recording and reconstruction of the video data in combination with the appropriately adapted audio data, the audio data will never been recorded or reproduced with deviation to the corresponding video data. Therefore, preparation of data file and cumbersome simulated reconstruction becomes unnecessary any more.
Furthermore, accord-ing to the present invention, by detecting speed difference between video decoding and reconstruction and audio decoding and reconstruction, the videodata decodingand reconstructing timing is controlled with reference to the operation speed of the audio decoding and reconstructing operation speed for correction of ' ~ 21~7727 synchronization, real time synchronization of the video reconstruction and audio reconstruction can be achieved so that disturbance for the audio reconstruction, such as interruption of voice in sequential reconstruction of the voice or tone variation can be successfully avoided.
Although the invention has been illustrated and described with respect to exemplary embodiment thereof, it should be understood by those skllled in the art that the foregoing and various other changes, omissions and additions may be made therein and thereto, without departing from the spirit and scope of the present invention. Therefore, the present invention should not be understood as limited to the specific embodiment set out above but to include all possible embodiments which can be embodies within ascopeencompassed and equivalents thereof with respect to the feature set out in the appended claims.

Claims (13)

THE EMBODIMENTS OF THE INVENTION IN WHICH AN EXCLUSIVE
PROPERTY OR PRIVILEGE IS CLAIMED ARE DEFINED AS FOLLOWS:
1. A synchronous compression system for compressing input video data and audio data with establishing synchronization therebetween, comprising:
video data compressing means for compressing each of first to nth blocks of video data where n is an integer greater than or equal to two in order of the first block to the nth block;
audio data compressing means for compressing input audio data for one frame;
interruption control means for providing control by interrupting video data compression when compression of video data is delayed with respect to compression of audio data with reference to frame numbers of those frames currently compressed in said video data compressing means and said audio data compressing means.
2. A synchronous compression system as set forth in claim 1, wherein said interruption control means performs interruption control per block.
3. A synchronous compression system as set forth in claim 1, wherein said video data compression means divides each block into first to mth frequency band components, in which m is an integer greater than or equal to two, and compression for each block is performed by scanning first to nth blocks in the first frequency band component in order and subsequently scanning respective first to nth blocks in second and subsequent frequency band components in order.
4. A synchronous compression system as set forth in claim 1, wherein said video data compression means divides each block into first to kth chrominance components, in which k is an integer greater than or equal to two, and compression of each block is performed by scanning the first to nth blocks of the first chrominance component in order and subsequently scanning first to nth blocks of respective of second and subsequent chrominance components in order.
5. A synchronous reconstruction system for synchronously reconstructing input video data and audio data respectively compressed per frame, comprising:
video data reconstructing means for scanning first to nth blocks of one frame input video data in order, which each one frame of video data is divided into first to nth blocks, audio data reconstructing means for reconstructing input audio data for one frame;
interruption control means for controlling to interrupt video data reconstruction when video data reconstruction process is delayed with respect to audio data reconstruction process with reference to frame numbers of those currently compressed in said video data reconstructing means and said audio data reconstructing means.
6. A synchronous reconstruction system as set forth in claim 5, wherein said interruption control means performs interruption control per block.
7. A synchronous reconstruction system as set forth in claim 5, wherein said video data reconstruction means divides each block into first to mth frequency band components, in which m is an integer greater than or equal to two, and reconstruction for each block is performed by scanning first to nth blocks in the first frequency band component in order and subsequently scanning respective first to nth blocks in second and subsequent frequency band components in order.
8. A synchronous reconstruction system as set forth in claim 5, wherein said video data reconstruction means divides each block into first to kth chrominance components, in which k is an integer greater than or equal to two, and reconstruction of each block is performed by scanning the first to nth blocks of the first chrominance component in order and subsequently scanning first to nth blocks of respective of second and subsequent chrominance components in order.
9. A synchronous reconstruction system for synchronously decoding and reconstructing digitally coded video data and audio data, comprising:

storage means for preliminarily storing digitally coded video and audio data;
video data reconstructing means for decoding and reconstructing digitally coded video data per predetermined data amount;
audio data reconstructing means for decoding and reconstructing audio data;
transfer control means for reading out digitally coded audio data from said storage means and transferring the read out digitally coded audio data to respective of said video data reconstructing means and said audio data reconstructing means; and control means responsive to completion of decoding and reconstructing process of both of said video data reconstructing means and said audio data reconstructing means for controlling modification of initiation timing of said video data reconstructing means and image display period after video reconstruction process.
10. A synchronous reconstruction system as set forth in claim 9, wherein said control means comprises:
time measuring means in response to completion of decoding and reconstructing of one of said video data reconstructing means and said audio data reconstructing means for initiating measurement of an elapsed tome and terminating said measurement in response to completion of decoding and reconstructing process of both of said video data reconstructing means and said audio data reconstructing means;
means responsive to termination of measurement of the elapsed period by said time measuring means, for deriving a process speed difference between said video data reconstruction process and said audio data reconstruction process with reference to the measured period; and means for controlling modification of an initiation timing of said video data reconstructing means or display period of video data after video data reconstructing process depending upon the processing speed difference.
11. A synchronous reconstruction system as set forth in claim 10, wherein said control means detects completion of process for a given amount of audio data within a predetermined unit period said audio data reconstructing means and detects completion of process for a given video data amount to be reproduced by said video data reconstructing means.
12. A synchronous reconstruction system as set forth in claim 11, wherein said control means controls to expand or shorten the display period for the video data per data unit following to video data amount to be reproduced within said unit period when the processing speed difference is smaller than or equal to a given value.
13. A synchronous reconstruction system as set forth in claim 11, wherein said control means controls thinning or holding of video data per data unit following to video data amount to be reproduced within said unit period when the processing speed difference is greater than a given value.
CA002107727A 1992-10-07 1993-10-05 Synchronous compression and reconstruction system Expired - Lifetime CA2107727C (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP293716/92 1992-10-07
JP4293716A JPH0822065B2 (en) 1992-10-07 1992-10-07 Image and audio information compression / reproduction system
JP1603593A JP2500579B2 (en) 1993-02-03 1993-02-03 Video and audio synchronization system
JP16035/93 1993-02-03

Publications (2)

Publication Number Publication Date
CA2107727A1 CA2107727A1 (en) 1994-04-08
CA2107727C true CA2107727C (en) 1999-06-01

Family

ID=26352275

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002107727A Expired - Lifetime CA2107727C (en) 1992-10-07 1993-10-05 Synchronous compression and reconstruction system

Country Status (5)

Country Link
US (1) US5508816A (en)
EP (1) EP0591944B1 (en)
AU (2) AU668762B2 (en)
CA (1) CA2107727C (en)
DE (1) DE69327563T2 (en)

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5850500A (en) 1995-06-28 1998-12-15 Kabushiki Kaisha Toshiba Recording medium comprising a plurality of different languages which are selectable independently of each other
DE69318038T2 (en) * 1993-10-29 1998-10-29 Toshiba Kawasaki Kk RECORDING MEDIUM THAT CAN PROCESS DIFFERENT LANGUAGES AND PLAYBACK DEVICE
JP3568984B2 (en) * 1994-06-20 2004-09-22 株式会社日立製作所 Information reproducing method, reproducing apparatus, output method and output apparatus
JPH0887577A (en) * 1994-09-19 1996-04-02 Olympus Optical Co Ltd Information recording medium and information reproducing device
US5598352A (en) * 1994-09-30 1997-01-28 Cirrus Logic, Inc. Method and apparatus for audio and video synchronizing in MPEG playback systems
US5594660A (en) * 1994-09-30 1997-01-14 Cirrus Logic, Inc. Programmable audio-video synchronization method and apparatus for multimedia systems
US6330644B1 (en) 1994-10-27 2001-12-11 Canon Kabushiki Kaisha Signal processor with a plurality of kinds of processors and a shared memory accessed through a versatile control means
US5721720A (en) * 1994-12-28 1998-02-24 Kabushiki Kaisha Toshiba Optical recording medium recording pixel data as a compressed unit data block
US6009202A (en) * 1994-12-28 1999-12-28 Kabushiki Kaisha Toshiba Image information encoding/decoding system
CA2168327C (en) * 1995-01-30 2000-04-11 Shinichi Kikuchi A recording medium on which a data containing navigation data is recorded, a method and apparatus for reproducing a data according to navigationdata, a method and apparatus for recording a data containing navigation data on a recording medium.
CN1086498C (en) * 1995-02-22 2002-06-19 株式会社东芝 Information recording method, recording media, information reproducing method and information reproducing device
EP0735782A3 (en) * 1995-03-31 1998-01-07 Kabushiki Kaisha Toshiba Motion picture play apparatus capable of returning from trick play to normal play
US5751694A (en) * 1995-05-22 1998-05-12 Sony Corporation Methods and apparatus for synchronizing temporally related data streams
US5818967A (en) * 1995-06-12 1998-10-06 S3, Incorporated Video decoder engine
JP3692164B2 (en) * 1995-06-20 2005-09-07 ユナイテッド・モジュール・コーポレーション MPEG decoder
DE19524808A1 (en) * 1995-07-07 1997-01-09 Thomson Brandt Gmbh Process, encoder and decoder for resynchronization to a faulty data stream
US5617502A (en) * 1996-03-22 1997-04-01 Cirrus Logic, Inc. System and method synchronizing audio and video digital data signals during playback
CN1150456C (en) * 1996-10-15 2004-05-19 松下电器产业株式会社 Video and audio coding method, coding apparatus, and coding program recording medium
JP3063824B2 (en) * 1996-10-29 2000-07-12 日本電気株式会社 Audio / video synchronous playback device
JP3106987B2 (en) * 1997-01-09 2000-11-06 日本電気株式会社 Audio / video synchronous playback device
JP3000949B2 (en) * 1997-02-12 2000-01-17 日本電気株式会社 Audio / video synchronous compression / playback apparatus and method
CN1278386A (en) * 1997-09-05 2000-12-27 松下电器产业株式会社 Decoding method and recording medium carrying recorded decoding program
JP3063838B2 (en) * 1997-10-02 2000-07-12 日本電気株式会社 Audio / video synchronous playback apparatus and method
DE19754296A1 (en) 1997-12-08 1999-06-10 Thomson Brandt Gmbh Synchronization device
JP3578069B2 (en) * 2000-09-13 2004-10-20 日本電気株式会社 Long-term image / sound compression apparatus and method
US8228849B2 (en) * 2002-07-15 2012-07-24 Broadcom Corporation Communication gateway supporting WLAN communications in multiple communication protocols and in multiple frequency bands
US7433824B2 (en) * 2002-09-04 2008-10-07 Microsoft Corporation Entropy coding by adapting coding between level and run-length/level modes
DE60330198D1 (en) 2002-09-04 2009-12-31 Microsoft Corp Entropic coding by adapting the coding mode between level and run length level mode
US8290353B2 (en) * 2003-02-27 2012-10-16 Panasonic Corporation Data processing device and method
US7724827B2 (en) * 2003-09-07 2010-05-25 Microsoft Corporation Multi-layer run level encoding and decoding
US7684981B2 (en) * 2005-07-15 2010-03-23 Microsoft Corporation Prediction of spectral coefficients in waveform coding and decoding
US7693709B2 (en) * 2005-07-15 2010-04-06 Microsoft Corporation Reordering coefficients for waveform coding or decoding
US7599840B2 (en) * 2005-07-15 2009-10-06 Microsoft Corporation Selectively using multiple entropy models in adaptive coding and decoding
US7933337B2 (en) * 2005-08-12 2011-04-26 Microsoft Corporation Prediction of transform coefficients for image compression
US7565018B2 (en) * 2005-08-12 2009-07-21 Microsoft Corporation Adaptive coding and decoding of wide-range coefficients
US8184710B2 (en) * 2007-02-21 2012-05-22 Microsoft Corporation Adaptive truncation of transform coefficient data in a transform-based digital media codec
CN101076099B (en) * 2007-06-14 2010-06-09 北京中星微电子有限公司 Method and device for controlling video record and synchronized-controlling unit
US8179974B2 (en) 2008-05-02 2012-05-15 Microsoft Corporation Multi-level representation of reordered transform coefficients
JP5212473B2 (en) * 2008-06-26 2013-06-19 富士通セミコンダクター株式会社 VIDEO / AUDIO DATA OUTPUT DEVICE AND VIDEO / AUDIO DATA OUTPUT METHOD
US8406307B2 (en) 2008-08-22 2013-03-26 Microsoft Corporation Entropy coding/decoding of hierarchically organized data
US20170264942A1 (en) * 2016-03-11 2017-09-14 Mediatek Inc. Method and Apparatus for Aligning Multiple Audio and Video Tracks for 360-Degree Reconstruction
CN111623986A (en) * 2020-05-19 2020-09-04 安徽智寰科技有限公司 Signal feature extraction method and system based on synchronous compression transformation and time-frequency matching

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS57152786A (en) * 1981-03-17 1982-09-21 Pioneer Electronic Corp Video format signal recording and reproducing system
USRE33535E (en) * 1985-09-16 1991-02-12 Audio to video timing equalizer method and apparatus
JP2599389B2 (en) * 1987-06-17 1997-04-09 松下電器産業株式会社 Information editing device
JP2545083B2 (en) * 1987-04-23 1996-10-16 松下電器産業株式会社 Information editing device
DE69032361T2 (en) * 1989-03-16 1998-10-29 Fujitsu Ltd VIDEO / AUDIO MULTIPLEX TRANSMISSION SYSTEM
EP0471118B1 (en) * 1990-08-13 1995-12-20 Matsushita Electric Industrial Co., Ltd. A video signal digital recording and reproducing apparatus
US5138440A (en) * 1990-10-29 1992-08-11 General Instrument Corporation Method and apparatus for communicating a plurality of asynchronous signals over a digital communication path
US5253275A (en) * 1991-01-07 1993-10-12 H. Lee Browne Audio and video transmission and receiving system
JPH0574110A (en) * 1991-09-12 1993-03-26 Sony Corp Digital vtr

Also Published As

Publication number Publication date
AU679872B2 (en) 1997-07-10
US5508816A (en) 1996-04-16
EP0591944A2 (en) 1994-04-13
DE69327563T2 (en) 2000-08-24
EP0591944B1 (en) 2000-01-12
AU4810696A (en) 1996-05-23
AU668762B2 (en) 1996-05-16
CA2107727A1 (en) 1994-04-08
AU4883093A (en) 1994-04-21
EP0591944A3 (en) 1995-09-06
DE69327563D1 (en) 2000-02-17

Similar Documents

Publication Publication Date Title
CA2107727C (en) Synchronous compression and reconstruction system
JP3106987B2 (en) Audio / video synchronous playback device
US6339760B1 (en) Method and system for synchronization of decoded audio and video by adding dummy data to compressed audio data
CA2219154C (en) An apparatus for a synchronized playback of audio-video signals
EP0920219B1 (en) Audio/video synchronous reproducer enabling accurate synchronization between audio and video data and a method of audio/video synchronous reproduction
JP2802694B2 (en) Image reproducing apparatus and method
KR100847534B1 (en) Apparatus and method for determining rendering duration of video frame
JPH09161456A (en) Data processing apparatus and method therefor
US6195503B1 (en) Image and sound recording/reproducing apparatus
JP2500579B2 (en) Video and audio synchronization system
JP2910575B2 (en) Audio signal processing circuit and image / audio decoder
JP3159098B2 (en) Synchronized playback device for image and audio
EP1312076B1 (en) Audio signal transmission
JP3170317B2 (en) Image processing device
JPH06337784A (en) Information processor
JP2002152699A (en) Method and system for synchronously reproducing moving picture
JPH06121276A (en) Compression reproduction system for picture and audio information
JP2596349B2 (en) Compression rate adjustment method
JP2835293B2 (en) Image transmission device
JP3165661B2 (en) Audio synchronized playback device
JPH05130422A (en) Picture data decoder
JPH0787324A (en) Animation/voice synchronization processing system
JPH1093959A (en) Mobile image and sound compression system
JPH06350592A (en) Coded signal decoder
JPH06189283A (en) Transmitting device for digital picture signal

Legal Events

Date Code Title Description
EEER Examination request
MKEX Expiry

Effective date: 20131007

MKEX Expiry

Effective date: 20131007