WO2014192133A1 - 情報処理装置、情報処理方法、及び情報処理プログラム - Google Patents
情報処理装置、情報処理方法、及び情報処理プログラム Download PDFInfo
- Publication number
- WO2014192133A1 WO2014192133A1 PCT/JP2013/065186 JP2013065186W WO2014192133A1 WO 2014192133 A1 WO2014192133 A1 WO 2014192133A1 JP 2013065186 W JP2013065186 W JP 2013065186W WO 2014192133 A1 WO2014192133 A1 WO 2014192133A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- output
- information processing
- imaged
- processing apparatus
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/147—Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
Definitions
- the present invention relates to a technical field such as a system for transmitting a user image captured by a camera to a terminal of a communication partner of the user.
- a videophone system that makes a call by transmitting video captured by a camera together with audio to a communication partner.
- video on the transmission side is constantly transmitted to the reception side.
- the user on the transmission side is not suitable for making a call with the user on the reception side.
- a situation may be assumed in which a user on the transmission side suddenly leaves the seat or a ringing tone sounds on the user's mobile phone. In such a case, there is a problem of causing distrust or discomfort to the receiving user who is watching the video of the transmitting user.
- Patent Document 1 is configured in advance so that camera image transmission on / off can be set in the storage unit, and when set to on, when the video from the camera unit is set to off, A system is disclosed that can automatically switch and transmit images or videos (illustrations, self-portraits, and other arbitrary videos) stored in advance in a storage unit.
- Patent Document 1 simply by switching an image or a video stored in the storage unit in advance before starting a videophone currently being performed to a video from the camera unit and transmitting it, the transmission side Since the state of the user is not transmitted to the user on the receiving side, it is difficult to avoid causing anxiety and the like to the user on the receiving side.
- the present invention has been made paying attention to the above points as an example of the problem, and is capable of avoiding anxiety and the like from being caused to the user on the receiving side. It is an object to provide a method and an information processing program.
- an invention according to claim 1 is an information processing apparatus that outputs an image captured by an image capturing unit in image communication performed using an image capturing unit and an image display device.
- An acquisition unit that acquires the video imaged by the imaging unit, a storage control unit that stores at least a part of the video image acquired by the acquisition unit in the storage unit, and a video that is currently stored in the storage unit
- a determination means for determining a video of a predetermined section after the start of the video communication being started as a replacement video, and a trigger for satisfying a predetermined start condition during the output of the video imaged by the imaging means. Instead, output means for outputting the replacement video is provided.
- the video in the predetermined section after the start of the current video communication is output as the replacement video, so that the user on the video receiving side is anxious. It is possible to avoid generating a feeling or the like.
- the information processing apparatus changes a subject to be imaged that appears in the video acquired by the acquisition unit during the output of the replacement video.
- First detecting means for detecting the image from the video wherein the storage control means stores at least the video of the section indicating the process of change detected by the first detecting means in the storage means, and the output means comprises: A video of a section indicating the process of the change is output following the replacement video.
- the images before and after the change of the person to be imaged can be connected in a natural manner, it is possible to avoid causing anxiety and the like to the user on the image receiving side.
- the information processing device acquires information indicating an action of a person who is viewing the video displayed by the video display device.
- the information acquisition means is further provided, and the output means represents the action of the person to be imaged in accordance with the action indicated by the information acquired by the information acquisition means instead of the video imaged by the imaging means.
- a replacement video is output.
- the user on the video receiving side can recognize that the user on the video transmitting side is performing an operation suitable for the user on the video receiving side. Can be avoided.
- the information processing apparatus performs a predetermined operation of the person to be imaged that appears in the video acquired by the acquisition unit.
- Second detection means for detecting the image from the video, and the output means is triggered by the detection of the predetermined action by the second detection means during the output of the video imaged by the imaging means.
- the replacement video is output instead of the video imaged by the imaging means.
- the replacement video can be output instead of the video imaged by the imaging means only by the user on the video transmission side performing a predetermined operation.
- the information processing apparatus is a place where a subject to be imaged that appears in the video acquired by the acquisition unit exists. Further comprising first determination means for determining whether or not the sound emitted by the sound belongs to a predetermined sound type, wherein the output means is configured to output the first determination means during the output of the video imaged by the imaging means. The replacement video is output in place of the video imaged by the imaging means, triggered by determining that the sound belongs to the predetermined sound type.
- the replacement video can be output in place of the video imaged by the imaging means, so that the user on the video transmission side is saved. be able to.
- a sixth aspect of the present invention is the information processing apparatus according to any one of the first to fifth aspects, wherein the information processing apparatus is a situation in which a person to be imaged appears in the video acquired by the acquisition unit.
- a second determination unit that determines whether or not the imaging subject is in a state in which the second determination unit speaks during the output of the video imaged by the imaging unit. If it is determined, the replacement video indicating a communication failure is output.
- the output unit is configured to output all of the determination units determined during the replacement video output.
- the replacement video that has already been output is repeatedly output.
- the present invention it is possible to continue the transmission of the replacement video even when the replacement video output time is long and the replacement video is insufficient.
- the output unit converts a partial area of the video captured by the imaging unit to the replacement video. An image is output in place of a partial area including the person to be imaged that appears in
- this target when there is a target to be shown to the user on the video receiving side in an area other than a partial area of the video captured by the imaging means, this target can be shown to the user on the video receiving side.
- the invention according to claim 9 is an information processing method executed by a computer that outputs an image picked up by the image pickup means in a video communication performed using the image pickup means and the video display device.
- a step of acquiring the captured video, a step of storing at least a part of the acquired video in the storage unit, and a predetermined time after the start of video communication currently performed from the video stored in the storage unit A step of determining a video of a section as a replacement video, and a step of outputting the replacement video in place of the video triggered by satisfying a predetermined start condition during the output of the video captured by the imaging means , Including.
- This information processing program is recorded on a computer-readable recording medium.
- an image of a predetermined section after the start of currently performed video communication is output as a replacement video instead of the video captured by the imaging unit, so that the user on the video receiving side is uneasy. It is possible to avoid generating a feeling or the like.
- FIG. 1 It is a figure showing an example of outline composition of videophone system S concerning this embodiment.
- A is a figure which shows the example of a schematic structure of the communication terminal Tn
- B is a figure which shows an example of the functional block in the control part 9.
- FIG. It is a flowchart which shows the communication process in the control part 9 of communication terminal T1.
- FIG. 1 is a diagram showing a schematic configuration example of a videophone system S according to the present embodiment.
- the communication terminal Tn is an example of an information processing apparatus of the present invention.
- Each of the communication terminals Tn can be connected to the network NW.
- the network NW is constructed by, for example, the Internet, a mobile communication network (including a base station), and a gateway.
- the communication terminal Tn can communicate with another communication terminal Tn using, for example, the IP (Internet Protocol) protocol via the network NW.
- IP Internet Protocol
- a personal computer, a mobile phone, and a mobile terminal such as a smartphone can be applied to the communication terminal Tn.
- FIG. 2A is a diagram illustrating a schematic configuration example of the communication terminal Tn.
- the communication terminal Tn includes a camera 1 (an example of an imaging unit), a microphone 2, a display unit 3, an audio output unit 4, a storage unit 5, an operation unit 6, a communication unit 7, and input / output.
- An interface unit 8 and a control unit 9 are provided.
- the camera 1, microphone 2, display unit 3, audio output unit 4, storage unit 5, operation unit 6, and communication unit 7 are connected to the input / output interface unit 8.
- the input / output interface unit 8 and the control unit 9 are connected via a system bus 10.
- the camera 1 captures an image of the user of the communication terminal Tn as an imaging subject when video communication is performed between the users of the communication terminal Tn via, for example, a network NW in a videophone conference.
- the camera 1 may not be mounted on the communication terminal Tn and may be an external camera connected to the communication terminal Tn.
- the microphone 2 collects user's voice and the like in synchronization with imaging by the camera 1.
- the display unit 3 includes a display and displays a video on the display according to the video signal output from the control unit 9.
- This video signal is, for example, a video signal representing a video of a user of the communication terminal Tn that is a communication partner.
- the audio output unit 4 includes a speaker, and outputs audio from the speaker to the outside according to the audio signal output from the control unit 9.
- This voice signal includes, for example, the voice of the user of the communication terminal Tn of the communication partner.
- the storage unit 5 is composed of, for example, a flash memory or a hard disk drive, and stores various programs such as an operating system and a videophone program (including the information processing program of the present invention).
- the information processing program of the present invention is a program that causes the control unit 9 to execute each step of the present invention.
- the information processing program of the present invention may be an application program provided separately from the videophone program.
- the videophone program may be downloaded from a predetermined server or the like via the network NW, or may be recorded on a predetermined recording medium and read.
- the operation unit 6 includes a plurality of operation buttons for receiving an instruction from the user, and outputs an instruction signal corresponding to the operation button pressed by the user to the control unit 9.
- the communication unit 7 is connected to the network NW and controls the normal state.
- the control unit 9 includes a CPU (Central Processing Unit) 91, a ROM (Read Only Memory) 92, a RAM (Random Access Memory) 93, and the like.
- FIG. 2B is a diagram illustrating an example of functional blocks in the control unit 9.
- the control unit 9 as a computer executes processing to be described later according to a videophone program stored in the storage unit 5.
- the control unit 9 includes a video acquisition unit 9a, a video storage control unit 9b, a video determination unit 9c, an information acquisition unit 9d, an audio acquisition unit 9e, and an output control unit.
- storage part 5 comprises the memory
- the RAM 93 or the storage unit 5 is allocated with a video storage area for temporarily storing the video captured by the camera 1.
- the video acquisition unit 9a is an example of an acquisition unit in the present invention.
- the video storage control unit 9b is an example of the storage control means in the present invention.
- the video determination unit 9c is an example of a determination unit in the present invention.
- the information acquisition unit 9d is an example of an information acquisition unit in the present invention.
- the output control unit 9f is an example of output means in the present invention.
- the motion detector 9h is an example of a second detector in the present invention.
- the sound type determination unit 9i is an example of a first determination unit in the present invention.
- the change detection unit 9j is an example of a first detection unit in the present invention.
- the speech determination unit 9k is an example of a second determination unit in the present invention.
- the video acquisition unit 9a acquires the video captured by the camera 1 as a video signal.
- the video storage control unit 9b stores at least a part of the video (hereinafter referred to as “real-time video”) acquired by the video acquisition unit 9a in a video storage area (hereinafter referred to as “real-time video storage area”). That is, the real-time video storage area does not have to store all the real-time videos acquired by the video acquisition unit 9a.
- the video storage control unit 9b may be configured to store only a video showing the person to be imaged in the real-time video storage area.
- the video determining unit 9c determines a video in a predetermined section after the start of the current video communication from the video stored in the real-time video storage area as a replacement video (hereinafter referred to as “dummy video”). To do. For example, from a video stored in the real-time video storage area, a video in a predetermined section in which the person to be imaged appears is determined as a dummy video.
- the predetermined section in which the imaging target person appears may include a section that is outside the imaging range for several seconds in order for the imaging target person to tie a shoelace, for example.
- the video determination unit 9c detects a person area from a plurality of image frames constituting the real-time video stored in the real-time video storage area, and tracks the detected person area in the plurality of image frames. To do. Then, the video determination unit 9c determines, as dummy videos, a plurality of image frames in a section in which a change in the tracked person area is equal to or less than a preset threshold from a plurality of image frames including the tracked person area. To do.
- the “threshold value” is set so that an image of a section in which the operation of the person to be imaged does not change greatly is determined.
- the dummy video includes the video of the section where the imaging subject is stationary, but the video of the section where the imaging subject is moving moderately may be natural . Therefore, in order to determine the video of the section in which the person to be imaged is moving moderately as a dummy video, the change in the tracked person area is equal to or higher than the first threshold set in advance and the second threshold (from the first threshold). A plurality of image frames in a section that is less than or equal to (large) may be determined as dummy video.
- the determined dummy video is temporarily stored in the dummy video storage area that is distinguished from the real-time video storage area.
- various known methods such as a human face detection method and a human tracking method using an optical flow can be used for detection of the person region, and thus detailed description thereof is omitted.
- the information acquisition unit 9d acquires information indicating a predetermined operation (hereinafter referred to as “communication partner side operation”) of a person (a user on the video receiving side) who is viewing the real-time video displayed by the communication terminal Tn of the communication partner. To do.
- Information indicating the communication partner side operation is received from, for example, the communication unit 7 and acquired from the video signal input by the input control unit 9g.
- the communication partner side operation include a laughing operation and an applause operation. Note that the laughing state can be detected by a known face recognition method.
- the voice acquisition unit 9e acquires the voice collected by the microphone 2 as a voice signal.
- the collected sound includes sound (hereinafter referred to as “external sound”) emitted in a place where the imaging target person exists, in addition to the voice of the imaging target person.
- This external sound includes, for example, a chime, a ringtone of a mobile phone, and the like.
- the output control unit 9f outputs the real-time video acquired by the video acquisition unit 9a to the communication unit 7.
- the output control unit 9f outputs the real-time video acquired by the video acquisition unit 9a to the videophone program.
- the real-time video output from the output control unit 9f is transmitted to one or a plurality of communication terminals Tn (an example of a video display device) as a communication partner via the communication unit 7 and the network NW.
- the output control unit 9f is triggered by satisfying a predetermined start condition (hereinafter referred to as “dummy video transmission start condition”) described later during the output of the real-time video as a trigger, and the video determination unit
- the dummy video determined by 9c is output to the communication unit 7 or the videophone program.
- the dummy video output from the output control unit 9f is transmitted to one or a plurality of communication terminals Tn as communication partners via the communication unit 7 and the network NW.
- the output control unit 9f may be configured to output the real-time video in which a partial area of the real-time video is replaced with a partial area including the person to be imaged that appears in the dummy video.
- the cut out area is a part of the real-time video (an area having the same coordinate position as the cut out area).
- the output control unit 9f may be configured to output a dummy video that matches the communication partner side operation indicated by the information acquired by the information acquisition unit 9d, instead of the real-time video.
- the user on the video receiving side recognizes that the user on the video transmitting side is operating in accordance with the user on the video receiving side. For example, if the communication partner side operation is a laughing operation, a dummy video including the laughing operation is transmitted. Alternatively, if the communication partner side operation is an applause operation, a dummy video including an applause operation is output.
- the video determination unit 9c extracts the dummy video corresponding to each action from the real-time video, and categorizes the extracted dummy video by, for example, each action (laughs, applause, etc.) and stores it in the dummy video storage area. Configured.
- the real-time video or the dummy video is encoded, and the encoded video signal is multiplexed with the audio signal acquired and encoded by the audio acquisition unit 9e and transmitted as stream data.
- the transmission of the dummy video ends when a dummy video transmission end condition described later is satisfied.
- the input control unit 9g inputs stream data transmitted from the communication terminal Tn of the communication partner via the network NW and received by the communication unit 7.
- the input control unit 9g separates and decodes the video signal and the audio signal from the input stream data, outputs the video signal to the display unit 3, and outputs the audio signal to the audio output unit 4.
- the motion detection unit 9h detects a predetermined motion of the person to be imaged that appears in the real-time video acquired by the video acquisition unit 9a from the real-time video. “Detecting a predetermined action of the person to be imaged” is an example of the above-described dummy video transmission start condition. For example, as the “predetermined motion”, a “slumber motion in which a person's head swings back and forth or right and left”, “a motion in which a person is asleep and eyes are closed for several tens of seconds”, “a motion in which a person stands up” A video representing “the action of facing the left or right of a person” is registered in advance.
- the motion detection unit 9h detects the motion if the similarity is greater than or equal to a preset threshold value compared to the registered motion and the human region detected from the real-time video. It should be noted that the state where the eyes are closed due to the falling asleep can be detected by a known face recognition method.
- the sound type determination unit 9i determines whether or not the external sound included in the sound acquired by the sound acquisition unit 9e belongs to a predetermined sound type.
- “External sound belongs to a predetermined sound type” is an example of the above-described dummy video transmission start condition.
- a frequency distribution representing “chimes” or “ringtones of mobile phones” is registered in advance.
- the sound type determination unit 9i compares the registered frequency distribution with the frequency distribution of the external sound included in the sound acquired by the sound acquisition unit 9e, for example, and the similarity is equal to or higher than a preset threshold value. For example, it is determined that the external sound belongs to a predetermined sound type.
- the output control unit 9f uses a dummy video instead of the real-time video as a trigger when a predetermined motion of the person to be imaged is detected by the motion detection unit 9h during the output of the real-time video acquired by the video acquisition unit 9a. Is output.
- the dummy video can be transmitted instead of the real-time video only by the user on the video transmission side performing a predetermined operation.
- the output control unit 9f uses the real-time video output by the video acquisition unit 9a as a trigger when the sound type determination unit 9i determines that the external sound belongs to a predetermined sound type.
- a dummy video is output instead of the video.
- the dummy video can be transmitted instead of the real-time video without the user on the video transmission side performing a special operation. Therefore, it is possible to save the trouble of the user on the video transmission side.
- the output control unit 9f is triggered in real time by a predetermined operation button (for example, a dummy video transmission start button) being pressed by the person to be imaged during the output of the real-time video acquired by the video acquisition unit 9a.
- a predetermined operation button for example, a dummy video transmission start button
- a dummy video may be output instead of the video.
- the operation button serving as a trigger may be configured to be arbitrarily set by the user from among a plurality of operation buttons.
- the change detection unit 9j detects a change in the person to be imaged that appears in the real-time video acquired by the video acquisition unit 9a from the real-time video during the output of the dummy video.
- Changes in the imaging target person include changes in the number of imaging target persons (for example, the number of imaging target persons imaged by the camera 1 changes from one to two) and changes in the operation of the imaging target person (for example, the imaging target person). Change due to the person removing his / her seat).
- the change detection unit 9j tracks a person area from a real-time video, monitors a change in the tracked person area from a plurality of image frames including the tracked person area, and changes the change.
- the video storage control unit 9b displays a video of a section indicating a process of change detected by the change detection unit 9j (for example, a process in which the person to be imaged captured by the camera 1 is changed from one to two).
- change process video is stored in a change assumption video storage area that is distinguished from the real-time video storage area and the dummy video storage area.
- the output control unit 9f outputs the change process video following the dummy video. Then, when all the change process videos stored in the change assumption video storage area are output, the output control unit 9f switches to the real time video and outputs it.
- the speech determination unit 9k determines whether or not the person to be imaged appears in the real-time video acquired by the video acquisition unit 9a. For example, when it is detected that the name of the imaging subject is called, it is determined that the imaging subject is in a situation to speak. That the name of the person to be imaged is called can be determined from the audio signal acquired by the audio acquisition unit 9e.
- the speech determination unit 9k receives a speech request from the administrator terminal so that the person to be imaged speaks. You may comprise so that it may determine with being. Note that the request for speech is transmitted from the terminal when, for example, the speech of the person to be imaged is designated with a button on the terminal of the administrator.
- the output control unit 9f outputs a dummy video indicating a communication failure for a predetermined time.
- the dummy video indicating a communication failure include a static dummy video and a dummy video with a reduced transmission bit rate.
- FIG. 3 is a flowchart showing communication processing in the control unit 9 of the communication terminal T1.
- the process illustrated in FIG. 3 is started, for example, when the communication terminal T1 starts communication with another communication terminal T2 or the like.
- step S1 the control unit 9 of the communication terminal T1 starts a video and audio acquisition process.
- video and audio acquisition processing real-time video is acquired by the video acquisition unit 9a, and audio is acquired by the audio acquisition unit 9e. Note that the video and audio acquisition processing is continued until, for example, the processing illustrated in FIG. 3 ends.
- the control unit 9 of the communication terminal T1 starts real-time video storage processing (step S2).
- the real-time video is stored in the real-time video storage area by the video storage control unit 9b.
- the real-time video storage process is continued until, for example, the process shown in FIG.
- the oldest real-time video stored in the real-time video storage area is It will be overwritten.
- the control unit 9 of the communication terminal T1 starts a dummy video determination process (step S3).
- the video determination unit 9c determines a dummy video in a predetermined section after the start of the current video communication from the real-time video.
- the determined dummy video is stored in the dummy video storage area.
- the dummy video determination process ends when, for example, the data amount of the dummy video stored in the dummy video storage area becomes equal to the capacity of the dummy video storage area.
- a new dummy video may be determined except for a section in which the dummy video is output instead of the real-time video.
- the dummy video is preferably determined from a real-time video acquired in the current session from the start to the end of communication with the communication terminal T1 and other communication terminals Tn. This is because even if the dummy video is determined from the real-time video acquired in the past sensation, the user's clothes and hairstyle may be different, so it may not be able to serve as a dummy video. .
- step S4 the control unit 9 of the communication terminal T1 starts real-time video transmission processing.
- stream data including real-time video and audio is transmitted to the communication terminal T2 and the like.
- the video and audio acquisition processing, real-time video storage processing, dummy video determination processing, and real-time video transmission processing are performed in parallel by, for example, a multitask function.
- the control unit 9 of the communication terminal T1 determines whether or not the dummy video transmission start condition is satisfied during transmission of the real-time video by the real-time video transmission process (during output by the output control unit 9f) (step S5). For example, while a video conference is being conducted through a plurality of communication terminals Tn, a situation has arisen in which the user of the communication terminal T1 needs to take a seat away due to poor physical condition or handle an emergency call. In this case, when the user presses the “dummy video transmission start button”, it is determined that the dummy video transmission start condition is satisfied.
- the sound type determination unit 9i determines that the external sound belongs to a predetermined sound type, for example, by sounding the entrance chime or the ring tone of the mobile phone. If it is determined, the dummy video transmission start condition is determined to be satisfied. In addition, when the user of the communication terminal T1 performs a large operation (for example, a rising operation or a left or right operation) and the operation detection unit 9h detects the operation of the person to be imaged, dummy video transmission starts. It is determined that the condition is satisfied. This is because the user is expected to temporarily leave the call soon.
- a large operation for example, a rising operation or a left or right operation
- step S6 it determines with the control part 9 of the communication terminal T1 not satisfy
- step S7 it determines with the control part 9 of communication terminal T1 satisfy
- step S6 it determines with the control part 9 of communication terminal T1 satisfy
- step S6 it determines with the control part 9 of communication terminal T1 satisfy
- step S7 it is determined whether or not there is a communication end instruction. For example, when it is determined that there is a communication end instruction from the user of the communication terminal T1 via the operation unit 6 (step S6: YES), the process illustrated in FIG. 3 ends. On the other hand, when it is determined that there is no communication end instruction (step S6: NO), the process returns to step S5.
- step S7 the control unit 9 of the communication terminal T1 starts a dummy video transmission process. That is, the dummy video transmission process is started with a condition that the dummy video transmission start condition is satisfied as a trigger.
- stream data including the dummy video is transmitted to the communication terminal T2 or the like instead of the real-time video.
- the voice (real-time voice) acquired by the voice acquisition unit 9e is cut (not transmitted). This is to prevent anxiety and the like from being given to a user such as the communication terminal T2 as a communication partner by transmitting audio not related to the dummy video.
- the dummy video transmission process may be configured to transmit the real-time video in which a partial area of the real-time video is replaced with a partial area including the person to be imaged that appears in the dummy video.
- control part 9 of communication terminal T1 transmits the dummy image
- You may comprise as follows. For example, if the communication partner side operation is a laughing operation, a dummy video including the laughing operation is transmitted. Alternatively, if the communication partner side operation is an applause operation, a dummy video including an applause operation is transmitted.
- the control unit 9 of the communication terminal T1 determines whether or not the change detection unit 9j detects a change in the person to be imaged that appears in the real-time video acquired by the video acquisition unit 9a during transmission of the dummy video (Ste S8). If a change in the person to be imaged appearing in the real-time video (for example, a change in the number of persons to be imaged imaged by the camera 1 from one to two) is detected (step S8: YES), the process proceeds to step S9. On the other hand, when a change in the person to be imaged that appears in the real-time video is not detected (step S8: NO), the process proceeds to step S10.
- step S9 the control unit 9 of the communication terminal T1 stores the change process video of the section indicating the detected change process in the change assumed video storage area, and proceeds to step S10.
- step S10 the control unit 9 of the communication terminal T1 uses the speech determination unit 9k to determine whether the subject to be imaged appears in the real-time video acquired by the video acquisition unit 9a during transmission of the dummy video. judge.
- step S10: YES the process proceeds to step S11. .
- step S10 determines that the person to be imaged appearing in the real-time video is not in a situation (step S10: NO).
- step S11 the control unit 9 of the communication terminal T1 transmits a dummy video indicating a communication failure to the communication terminal Tn of the communication partner for a predetermined time, and proceeds to step S12.
- the control unit 9 of the communication terminal T1 When the transmission of all the dummy videos determined by the video determination unit 9c and stored in the dummy video storage area is completed during transmission of the dummy video, the control unit 9 of the communication terminal T1 has already transmitted the dummy video. Is repeatedly transmitted to the communication terminal T2 or the like. That is, when there are not enough dummy videos stored in the dummy video storage area, the transmission is repeated so that the joints become smooth. As a result, the transmission time of the dummy image becomes longer, and the transmission of the dummy image can be continued even when the dummy image becomes insufficient.
- step S12 the control unit 9 of the communication terminal T1 determines whether or not the dummy video transmission end condition is satisfied during transmission of the dummy video. For example, when a predetermined operation button (for example, a dummy video transmission end button) is pressed by the user of the communication terminal T1, it is determined that the dummy video transmission end condition is satisfied. For example, if the agenda of the video conference is decided and the time at which the user of the communication terminal T1 has to speak is known in advance, the user operates the operation unit 6 to set a timer. Also good. In this case, when the time set by the user arrives, the control unit 9 of the communication terminal T1 determines that the dummy video transmission end condition is satisfied.
- a predetermined operation button for example, a dummy video transmission end button
- step S12 determines with not satisfy
- step S12: NO it progresses to step S13.
- step S12: YES if it is determined that the dummy video transmission end condition is satisfied (step S12: YES), the process proceeds to step S14.
- step S13 it is determined whether or not there is a communication end instruction. If it is determined that there is a communication end instruction (step S13: YES), the process shown in FIG. 3 ends. On the other hand, when it is determined that there is no instruction to end communication (step S13: NO), the process returns to step S8.
- step S14 the control unit 9 of the communication terminal T1 determines whether or not the change process video is stored in the change assumed video storage area in step S9. When it is determined that the change process video is stored in the change assumed video storage area (step S14: YES), the dummy video transmission process is terminated, and the process proceeds to step S15. On the other hand, when it is determined that the change process video is not stored in the change assumed video storage area (step S14: NO), the dummy video transmission process is terminated, and the process proceeds to step S16.
- step S15 the control unit 9 of the communication terminal T1 transmits the change process video stored in the change assumed video storage area to the communication terminal Tn of the communication partner, and proceeds to step S16 after the transmission is completed.
- step S16 the control unit 9 of the communication terminal T1 restarts the real-time video transmission process and returns to step S5.
- the communication terminal Tn is currently performed instead of the real-time video as a trigger when a predetermined start condition is satisfied during the output of the real-time video acquired from the camera 1. Since it is configured to output a dummy video of a predetermined section after the video communication is started, it is possible to transmit a video (dummy video) in which the user on the video transmission side participates in, for example, a conference, It is possible to avoid causing anxiety and the like to the user on the video receiving side.
- the communication terminal on the video transmission side and the video reception side You may comprise so that it may be equipped with the mediation apparatus (For example, the server connected to network NW, or a set top box) interposed between communication terminals.
- the mediation apparatus acquires the real-time video captured by the camera 1 from the communication terminal T1, stores the acquired real-time video in the video storage area, and transmits the real-time video to the communication terminal T2 or the like.
- the intermediary device extracts a dummy video of a predetermined section in which the person to be imaged appears from the video stored in the video storage area, and triggers to satisfy the above-described dummy video transmission start condition during the output of the real-time video. Then, instead of the real-time video, the extracted dummy video is transmitted to the communication terminal T2. In this case, information for determining whether or not the dummy video transmission start condition is satisfied is acquired from the communication terminal T1.
- the acquisition unit, the storage control unit, the determination unit, the output unit, and the like in the present invention may be provided in an external device that can be connected to the communication terminal Tn via an interface such as a USB.
- the present invention is applied to a videophone system, and it is assumed that a plurality of users make a two-way videophone.
- one-way communication such as e-learning
- the present invention can also be applied to a system.
Abstract
Description
[1.テレビ電話システムSの構成及び機能]
[2.テレビ電話システムSの動作]
2 マイクロフォン
3 表示部
4 音声出力部
5 記憶部
6 操作部
7 通信部
8 入出力インターフェース部
9 制御部
Tn 通信端末
S テレビ電話システム
Claims (10)
- 撮像手段と映像表示装置を用いて行われる映像通信において当該撮像手段により撮像された映像を出力する情報処理装置であって、
前記撮像手段により撮像された映像を取得する取得手段と、
前記取得手段により取得された映像の少なくとも一部を記憶手段に記憶させる記憶制御手段と、
前記記憶手段に記憶された映像から現在行われている映像通信が開始された以降の所定区間の映像を差替映像として決定する決定手段と、
前記撮像手段により撮像された映像の出力中に所定の開始条件を満たすことをトリガとして、当該映像に代えて前記差替映像を出力する出力手段と、
を備えることを特徴とする情報処理装置。 - 前記情報処理装置は、前記差替映像の出力中に、前記取得手段により取得された映像に表れる撮像対象者の変化を当該映像から検知する第1検知手段を更に備え、
前記記憶制御手段は、少なくとも前記第1検知手段により検知された変化の過程を示す区間の映像を記憶手段に記憶させ、
前記出力手段は、前記差替映像に続けて前記変化の過程を示す区間の映像を出力することを特徴とする請求項1に記載の情報処理装置。 - 前記情報処理装置は、前記映像表示装置により表示された前記映像を見ている者の動作を示す情報を取得する情報取得手段を更に備え、
前記出力手段は、前記撮像手段により撮像された映像に代えて、前記情報取得手段により取得された情報が示す動作に合った、前記撮像対象者の動作が表れる前記差替映像を出力することを特徴とする請求項1または2に記載の情報処理装置。 - 前記情報処理装置は、前記取得手段により取得された映像に表れる撮像対象者の所定の動作を当該映像から検知する第2検知手段を更に備え、
前記出力手段は、前記撮像手段により撮像された映像の出力中に、前記第2検知手段により前記所定の動作が検知されることをトリガとして、前記撮像手段により撮像された映像に代えて前記差替映像を出力することを特徴とする請求項1乃至3の何れか一項に記載の情報処理装置。 - 前記情報処理装置は、前記取得手段により取得された映像に表れる撮像対象者が存在する場所で発せられる音が、所定の音の種類に属するか否かを判定する第1判定手段を更に備え、
前記出力手段は、前記撮像手段により撮像された映像の出力中に、前記第1判定手段により前記所定の音の種類に属すると判定されることをトリガとして、前記撮像手段により撮像された映像に代えて前記差替映像を出力することを特徴とする請求項1乃至3の何れか一項に記載の情報処理装置。 - 前記情報処理装置は、前記取得手段により取得された映像に表れる撮像対象者が発言する状況にあるか否かを判定する第2判定手段を更に備え、
前記出力手段は、前記撮像手段により撮像された映像の出力中に、前記第2判定手段により前記撮像対象者が発言する状況にあると判定された場合、通信障害を示す前記差替映像を出力することを特徴とする請求項1乃至5の何れか一項に記載の情報処理装置。 - 前記出力手段は、前記差替映像の出力中に、前記決定手段により決定された全ての前記差替映像の出力が完了した場合、既に出力した当該差替映像を繰り返して出力することを特徴とする請求項1乃至6の何れか一項に記載の情報処理装置。
- 前記出力手段は、前記撮像手段により撮像された映像の一部分の領域を、前記差替映像に表れる当該撮像対象者を含む一部分の領域に代えた映像を出力することを特徴とする請求項1乃至7の何れか一項に記載の情報処理装置。
- 撮像手段と映像表示装置を用いて行われる映像通信において当該撮像手段により撮像された映像を出力するコンピュータにより実行される情報処理方法であって、
前記撮像手段により撮像された映像を取得するステップと、
取得された映像の少なくとも一部を記憶手段に記憶させるステップと、
前記記憶手段に記憶された映像から現在行われている映像通信が開始された以降の所定区間の映像を差替映像として決定するステップと、
前記撮像手段により撮像された映像の出力中に所定の開始条件を満たすことをトリガとして、当該映像に代えて前記差替映像を出力するステップと、
を含むことを特徴とする情報処理方法。 - 撮像手段と映像表示装置を用いて行われる映像通信において当該撮像手段により撮像された映像を出力するコンピュータに、
前記撮像手段により撮像された映像を取得するステップと、
取得された映像の少なくとも一部を記憶手段に記憶させるステップと、
前記記憶手段に記憶された映像から現在行われている映像通信が開始された以降の所定区間の映像を差替映像として決定するステップと、
前記撮像手段により撮像された映像の出力中に所定の開始条件を満たすことをトリガとして、当該映像に代えて前記差替映像を出力するステップと、
を実行させることを特徴とする情報処理プログラム。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2013/065186 WO2014192133A1 (ja) | 2013-05-31 | 2013-05-31 | 情報処理装置、情報処理方法、及び情報処理プログラム |
US14/348,996 US9277174B2 (en) | 2013-05-31 | 2013-05-31 | Information processing apparatus, information processing method and information processing program |
JP2013531035A JP5422085B1 (ja) | 2013-05-31 | 2013-05-31 | 情報処理装置、情報処理方法、及び情報処理プログラム |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2013/065186 WO2014192133A1 (ja) | 2013-05-31 | 2013-05-31 | 情報処理装置、情報処理方法、及び情報処理プログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014192133A1 true WO2014192133A1 (ja) | 2014-12-04 |
Family
ID=50287270
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/065186 WO2014192133A1 (ja) | 2013-05-31 | 2013-05-31 | 情報処理装置、情報処理方法、及び情報処理プログラム |
Country Status (3)
Country | Link |
---|---|
US (1) | US9277174B2 (ja) |
JP (1) | JP5422085B1 (ja) |
WO (1) | WO2014192133A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017204845A (ja) * | 2016-05-12 | 2017-11-16 | 富士ゼロックス株式会社 | ビデオを動画と置き替える装置、プログラム及び方法 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10462409B2 (en) * | 2007-12-28 | 2019-10-29 | Google Technology Holdings LLC | Method for collecting media associated with a mobile device |
US8314838B2 (en) * | 2007-12-28 | 2012-11-20 | Motorola Mobility Llc | System and method for collecting media associated with a mobile device |
EP2950285B1 (en) * | 2014-05-26 | 2016-10-05 | Axis AB | Automatic configuration of a replacement camera |
CN105338148B (zh) * | 2014-07-18 | 2018-11-06 | 华为技术有限公司 | 一种根据频域能量对音频信号进行检测的方法和装置 |
CN113824920A (zh) * | 2021-09-30 | 2021-12-21 | 联想(北京)有限公司 | 一种处理方法及装置 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004328069A (ja) * | 2003-04-21 | 2004-11-18 | Nec Corp | テレビ電話端末および画像生成方法 |
JP2007189624A (ja) * | 2006-01-16 | 2007-07-26 | Mitsubishi Electric Corp | テレビ電話端末 |
JP2011223323A (ja) * | 2010-04-09 | 2011-11-04 | Ntt Docomo Inc | 対話型サービス提供システム、対話型サービス提供装置、動画像提供装置、対話型サービス提供方法、動画像提供プログラム |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10504692A (ja) * | 1994-08-19 | 1998-05-06 | ライカ アクチエンゲゼルシャフト | 立体ビデオ画像を表示装置上に表示する方法及び装置 |
JP2003009104A (ja) | 2001-06-25 | 2003-01-10 | Matsushita Electric Ind Co Ltd | テレビ電話機能付き電話機 |
JP2006332759A (ja) * | 2005-05-23 | 2006-12-07 | Toshiba Corp | 電子機器、画像制御方法、及び画像制御用のプログラム |
JP5877351B2 (ja) * | 2010-12-15 | 2016-03-08 | パナソニックIpマネジメント株式会社 | 通信装置および通信方法 |
-
2013
- 2013-05-31 JP JP2013531035A patent/JP5422085B1/ja active Active
- 2013-05-31 WO PCT/JP2013/065186 patent/WO2014192133A1/ja active Application Filing
- 2013-05-31 US US14/348,996 patent/US9277174B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004328069A (ja) * | 2003-04-21 | 2004-11-18 | Nec Corp | テレビ電話端末および画像生成方法 |
JP2007189624A (ja) * | 2006-01-16 | 2007-07-26 | Mitsubishi Electric Corp | テレビ電話端末 |
JP2011223323A (ja) * | 2010-04-09 | 2011-11-04 | Ntt Docomo Inc | 対話型サービス提供システム、対話型サービス提供装置、動画像提供装置、対話型サービス提供方法、動画像提供プログラム |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017204845A (ja) * | 2016-05-12 | 2017-11-16 | 富士ゼロックス株式会社 | ビデオを動画と置き替える装置、プログラム及び方法 |
Also Published As
Publication number | Publication date |
---|---|
US20150256794A1 (en) | 2015-09-10 |
JP5422085B1 (ja) | 2014-02-19 |
US9277174B2 (en) | 2016-03-01 |
JPWO2014192133A1 (ja) | 2017-02-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5422085B1 (ja) | 情報処理装置、情報処理方法、及び情報処理プログラム | |
US8319818B2 (en) | Information processing device, information processing method and storage medium storing computer program | |
US20110164105A1 (en) | Automatic video stream selection | |
US8599236B2 (en) | Utilizing a video image from a video communication session as contact information | |
CN107040751A (zh) | 控制实时会议会话的方法、使计算机执行该方法的计算机程序产品以及相关通信系统 | |
JP6432177B2 (ja) | 対話型通信システム、端末装置およびプログラム | |
US10362173B2 (en) | Web real-time communication from an audiovisual file | |
WO2022135005A1 (zh) | 一种基于通话的屏幕共享方法、装置、设备及存储介质 | |
JP2010239393A (ja) | 通信制御装置、通信制御方法、通信制御プログラム | |
CN110392304A (zh) | 一种视频展示方法、装置、电子设备及存储介质 | |
JP5609431B2 (ja) | 音量制御装置、音量制御方法およびプログラム | |
JP2018515979A (ja) | 通信処理方法及び電子装置 | |
CN110536092A (zh) | 视频留言方法、装置、电子设备及存储介质 | |
JP2005303734A (ja) | 通信装置およびサーバシステム | |
JP2010154387A (ja) | 通信端末装置、通信制御方法、通信制御プログラム | |
JP6550951B2 (ja) | 端末、ビデオ会議システム、及びプログラム | |
CN115665504A (zh) | 事件识别方法、装置、电子设备及存储介质 | |
JP2003309829A (ja) | 携帯動画電話装置 | |
CN101888522A (zh) | 网络视频会议设备及其进行网络视频会议的方法 | |
CN104468933A (zh) | 利用移动网络实现单人和多人语音留言的方法及设备 | |
CN110830974A (zh) | 应用智能终端的紧急求救系统和方法 | |
JP2015115926A (ja) | 携帯端末装置、読唇通信方法、及びプログラム | |
JP6680031B2 (ja) | 会議用装置、送信方法、およびコンピュータプログラム | |
KR100643451B1 (ko) | 화상 데이터와 텍스트 데이터의 동기화 출력 기능을가지는 화상 단말기 및 그 방법 | |
JP7420078B2 (ja) | 情報処理装置及び情報処理方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2013531035 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14348996 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13885641 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13885641 Country of ref document: EP Kind code of ref document: A1 |