EP4004918A1 - Media client for recording and playing back interpretation - Google Patents
Media client for recording and playing back interpretationInfo
- Publication number
- EP4004918A1 EP4004918A1 EP20739675.5A EP20739675A EP4004918A1 EP 4004918 A1 EP4004918 A1 EP 4004918A1 EP 20739675 A EP20739675 A EP 20739675A EP 4004918 A1 EP4004918 A1 EP 4004918A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- segment
- source
- interpreted
- media
- time
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 claims description 41
- 238000004590 computer program Methods 0.000 claims description 6
- 238000012552 review Methods 0.000 description 11
- 239000012634 fragment Substances 0.000 description 6
- 238000013519 translation Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 5
- 238000004891 communication Methods 0.000 description 5
- 238000012549 training Methods 0.000 description 5
- 230000007246 mechanism Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000011156 evaluation Methods 0.000 description 3
- 239000003814 drug Substances 0.000 description 2
- 230000002996 emotional effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000000151 deposition Methods 0.000 description 1
- 230000008021 deposition Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/04—Electrically-operated educational appliances with audible presentation of the material to be studied
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/11—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
Definitions
- the present invention generally relates, amongst others, to a media client and a method for recording and playing back interpretation. More particularly, it relates to a media client and a method for recording a sequence of interpreted segments, wherein each interpreted segment comprises an interpretation into a second language of a source segment of a source media comprising a first language.
- Interpreting is a translational activity in which one produces a translation based on an exposure to an expression in a source language, wherein the translation and/or the source language can be for example spoken or can be for example sign language.
- the most common two modes of interpreting are simultaneous interpreting, which is done at the time of the exposure to the source language, and consecutive interpreting, which is done at breaks to this exposure. Both for simultaneous and consecutive interpretation, the interpreter plays a key role in the exchange between partners whose language and culture are different.
- Simultaneous interpreting is the oral translation of a speech as it unfolds. In other words, during simultaneous interpretation, the interpreter does not interrupt the speech of the talker. The interpreter listens through to a speech which he or she translates in real time while listening. The listeners can hear the interpreted speech, wherein the interpreted speech is overlaid or not with the speech of the talker. This type of interpretation is often used as its benefits in terms of time savings are indisputable. For example, simultaneous interpreting is tailor-fit for large seminars and conferences. The great advantage of simultaneous interpretation is indeed that efficiency is high as the interpretation happens in a few seconds or minutes, and that it does not affect or interrupt the talker's thinking.
- Consecutive interpreting allows the oral transposition of a speech into another language just after the speech has finished.
- the principle is as follows: as the speaker talks, the interpreter listens, for example takes notes, and then reproduces the speech in the target language.
- the main difference between simultaneous interpreting and consecutive interpreting lies in how the comprehension phase is separated from the interpretation playback phase. Indeed, the simultaneous interpreter has little time to reproduce and properly structure the speech in the target language. With consecutive interpretation, the interpreter has more time. This means that he or she has the possibility to move away from the original speech, avoid literal translation more easily and even summarise as well as cut out any undesirable interferences.
- Consecutive interpreting is therefore often preferred to simultaneous interpreting in the context of human resources meetings for multilingual staff, press and personal interviews, business meeting, court depositions, medical interviews and consultations, tours and travels, etc.
- Interpretation is a high-level language translation activity with the added difficulty that it must be performed according to strict time restrictions. Interpreting requires the interpreter to listen accurately to distinguish source language speech, to rapidly plan, express, monitor and correctly predict, understand, memorize and convert source language information on a particular thematic knowledge to a target language different from the source language. Such skills can be trained by practicing live interpretation, both for live interpreting as for consecutive interpreting. On-the-job trainings for interpreters also exist to allow them to gather knowledge in specific technical domains requiring the use of precise detailed technical terms, such as for example healthcare, medicine, engineering, architecture and construction, finance, literature, etc. For instance, for medical interpreting, the interpreter should have prior experience in a particular field. They should know and understand the specific terminology used in the specific branch of medicine and translate the words into the language that the listener could easily understand.
- a media client for recording a sequence of interpreted segments, wherein each interpreted segment comprises a consecutive interpretation into a second language of a source segment of a source media comprising a first language, wherein the media client is configured to play back a first source segment of the source media comprising the first language; and wherein the media client is further configured to:
- the data comprises the first point in time and of the second point in time or a duration of the first interpreted segment determined between the second point in time and the first point in time.
- the media client according to the present disclosure allows to prepare interpreting exercises and send them to interpreters, to record performances of interpreters for consecutive interpreting exercises, and to evaluate interpretations and add feedback for the interpreters in the recording. Teachers or peers can evaluate these recorded interpretations rapidly and easily.
- the media client according to the present disclosure evaluates not only the spoken interpretation, but also the way the interpreter presents the message. Indeed, by capturing not only for example the interpreter's microphone, but also for example the webcam of the interpreter, the performance can be properly evaluated by all means.
- the media client according to the present disclosure allows a more emotional and fluent delivery of the interpretation.
- the media client according to the present disclosure makes it easy for interpreters to check the quality of their interpretation as well as the conformity with the source media.
- An interpreter and/or a teacher can quickly jump to points in time of interest in the source media. This is done in a user-friendly and intuitive manner.
- the media client according to the present disclosure allows recording a sequence of interpreted segments of a source media. The first point in time and the second point in time are determined with respect to the source media. The steps of this method are repeated for all the interpreted segments, thereby recording a sequence of interpreted segments.
- a teacher may simultaneously supervise, compare and/or evaluate the work of a plurality of interpreters and/or student interpreters which all use the media client according to the present invention.
- a media client is for example a media player configured to record and play media files.
- a media client in the context of the present disclosure is a video player.
- a media client is for example any type of media player software or any type of application software for playing multimedia computer files like audio and video files.
- a media client is for example a media player comprised in a web browser.
- a media client is for example a media player comprised in a web application, or a desktop application. Examples of media clients that are suitable as a media client are desktop and laptop computers, smartphones, tablets, setup boxes and TVs.
- a media client may also refer to a media player application, or application, running on any of such devices.
- the media client according to the present disclosure therefore allows teachers and interpreters and student interpreters to enjoy the benefits of practicing interpretation outside of the interpreting lab, taking away all limitations of time and place during preparation, practice or evaluation. It becomes possible to train professional interpreting skills through the intuitive training method according to the present disclosure, and to capture performances of interpreters during classes, live conferences and examinations. Additionally, the media client according to the present disclosure allows interpreters to train using the same equipment they will be using in their professional environment. This way, the media client according to the present disclosure maximizes the learning effect and ease of use for the interpreters, the teachers and/or the students.
- source media being played is for example audio or audio in combination with ordered still pictures or frames that are decoded or decompressed and played one after the other within for example a media client.
- source media is for example audio, such as a source track, or audio combined with video.
- fragment “slice”,“segment”,“part” are used throughout the text below as synonyms.
- the recording of interpreted “fragment”, interpreted “slice”, interpreted “segment”, “part” correspond to the recording the media client according to the present disclosure of the audio and/or video interpretation of source segments of the source media by an interpreter using the media client according to the present disclosure.
- interpreted segment is for example audio, such as a track, or audio combined with video, corresponding to the interpretation of a source segment of the source media.
- the term“interpret” relates to the translation of what is being said or signed into another language.
- the term“data” refers to for example“metadata”, for example one or more metadata files.
- the term“duration” refers to the length in time of media.
- the expressions comprising the term“recording” preferably refer to the interpretation of a source media by a user of the media client, i.e. preferably refer to interpreted segments.
- the resulting sequence of interpreted segments consists of one audio and/or video file.
- Information on where individual interpreted segments are exactly positioned in the sequence of interpreted segments is for example saved in the data.
- This information for example consists of the start time in relation to the source media and the duration of each interpreted segments of the sequence of interpreted segments.
- the first point in time and the second point in time are defined with respect to a duration of the source media.
- the media client determines and stores a timestamp of the first point in time with respect to the source media and the media client further determines and stores a timestamp of the second point in time with respect to the source media.
- the media client determines and stores a timestamp of the first point in time with respect to the source media and further determines and stores a duration of an interpreted segment determined between the second point in time and the first point in time. All the recorded interpreted segments, together with the data and the source media can for example subsequently be saved into a project file, such as for example into an IQF project file.
- Source media played by the media client is for example audio or audio in combination with ordered still pictures or frames that are decoded or decompressed and played one after the other within the media client.
- source media is for example audio, such as a source track, or audio combined with video.
- the media client may be any device capable of receiving a digital representation of such media over a communication network and capable of decoding the representation into a sequence of frames that can be displayed on a screen to a user.
- Interpreter training requires practicing live interpretation of a source fragment, this fragment being audio only or audio combined with video.
- a source fragment of source media for example comprises audio or audio combined with video, wherein the audio comprises a source language which is to be interpreted, for example during an interpretation exercise, by for example an interpreter and/or a teacher and/or a student.
- This live interpretation of a source segment can be:
- An interpreter typically starts by listening to a part or source segment of the source media.
- the student can click on for example the‘Start Recording’ button displayed on the media client.
- the playback of the source track is then paused and the video of the source media is frozen, and the microphone, i.e. audio and optionally the webcam, i.e. video, start being recorded by the media client.
- the interpreter can visually see a waveform being inserted into the recording track on the media client, while a flat line indicating silence is being inserted to the waveform representation of the source media.
- the interpreter can continue the exercise by clicking on for example the‘Continue playback’ button displayed on the media client.
- the recording of the first interpreted segment then pauses and the playback of source media resumes where it left off at the start of the previous recording, thereby defining an end to the first interpreted segment.
- the sequence of interpreted segments is recorded continuously and the timestamps of the first point in time and of the second point in time are used to determine the time boundaries of the first interpreted segment.
- the sequence of interpreted segments is recorded continuously and the timestamp of the first point in time and a duration of the first interpreted segment between the timestamp of the second point in time and the first point in time are determined.
- This embodiment has the technical advantage that reactions and/or emotions of the interpreter are recorded while the interpreter is listening to and/or looking at the video of the source media. Data such as for example metadata containing the start time of the recording of the first interpreted segment in relation to the source media and its duration is generated.
- the interpreter wants to translate the next source segment, the interpreter can click on for example the ‘Continue recording’ button displayed on the media client. This pauses the playback of the source media and resumes recording of the sequence of interpreted segments. This sequence of actions continues until the exercise is finished, i.e. for example until the whole source media has been interpreted.
- a teacher or the interpreter can play back the source media alongside the sequence of interpreted segments.
- a waveform visualization of the source media is generated based on the binary audio information of the source media file.
- this waveform is then adjusted based on the data of that particular recording.
- a flat line, indicating silence is added on the waveform of the source media where a recording interpreted segment is located according to its start time and duration.
- a waveform visualization of the recorded sequence of interpreted segments is based on the binary audio information of its corresponding media file. This waveform is then adjusted based on the data and the interpreting mode.
- the interpreted segment When a position of a time mark coincides with an interpreted segment, defined by its start time and duration, the interpreted segment is played at a position relative to the overall length while the source media is paused. When the position of the time mark does not match one of the interpreted segments, the sequence of interpreted segments is paused, and the source media is played at a position relative to the overall length of the exercise.
- the media client displays the source media and the video of the recording of the sequence of interpreted segments next to each other.
- the playback of the video of the source media pauses to start the recording of an interpreted segment
- the video of the recording of the interpreted segment starts; and when the recording of the sequence of the interpreted segments is paused to interpret a further source segment of the source media, the video of the recording of the interpreted segment pauses and the playback of the video of the source media starts again.
- the project file with for example extension IQF, can optionally be an encrypted and compressed into for example a container file, and comprises:
- a source media for example a source media file
- - data for example a metadata file.
- a metadata file for example comprises one or more of the following:
- An interpreted segment has a start time and a duration with respect to the start time being relative to the source media.
- Alternative ways of describing an interpreted segment can be to save the start and stop time relative to the source media, to save the interval between the interpreted segments or to save times relative to the total duration of the exercise. In simultaneous mode there is only one recording of a sequence of interpreted segments for each recorded version.
- the media client is further configured to:
- the interpreter can start the recording at any time relative to the source track. When the recording of the interpretation of the source fragment / track is finished, the interpreter can play back the recorded interpretation together with the source track.
- the media client according to the present disclosure allows recording a sequence of interpreted segments of a source media. The first point in time and the second point in time are determined with respect to the source media. The steps of this method are repeated for all the interpreted segments, thereby recording a sequence of interpreted segments. The consecutive interpretation can be played back by the media client such that the playback plays source segments consecutively followed by the recording of their respective interpreted segments.
- the video of the source media is for example paused or frozen at the first point in time, thereby allowing the playback of the recording of the first interpreted segment for the duration of the first interpreted segment, and at the second point in time, the recording of the first interpreted segment is paused or frozen such that a second source segment of the source media starting from the first point in time can be played back.
- This ensures either a source segment or an interpreted segment is always played back and shows the user of the media client that the video is still relevant even though not actively being played.
- the audio recording of the first interpreted segment is paused, but the video is still being recorded.
- the media client allows the export of for example an IQF file or a zip file comprising for example an MP3 or an MP4 file comprising only the recording of the sequence of interpreted segments, such as for example the audio and/or the video track of the source media and of the sequence of interpreted segments consecutively or simultaneously playing, or the audio and/or the video track of only the sequence of interpreted segments, and/or further comprising the data.
- the recorded video and/or audio of the sequence of interpreted segments comprises several files.
- additional data is stored in the project file in the media client, such as for example references to the different recorded files of the same exercise, and possibly the lengths, in milliseconds / seconds / frames / ... , of these files.
- the media client for example comprises a playback unit configured to play back the first source segment of the source media comprising the first language.
- the playback unit may be further configured to pause the playback of the first source segment at the first point in time.
- the media client for example further comprises a recording unit configured to, at the first point in time, record a first interpreted segment comprising a live consecutive interpretation into the second language of the first source segment.
- the recording unit is for example further configured to define an end of the recording of the first interpreted segment at a second point in time and the playback unit is further configured to, at the second point in time, play back a second source segment of the source media starting from the first point in time referenced with respect to the source media.
- the media client for example further comprises a metadata generator configured to generate metadata for the interpreted segments.
- the metadata generator is configured to determine a duration of the first interpreted segment between the second point in time and the first point in time and store data on the first interpreted segment, wherein the data comprises the first point in time and the duration of the first interpreted segment.
- the media client is further configured to pause the playback of the first source segment and to start recording the first interpreted segment simultaneously.
- the media client when the media client is further configured to play back the second source segment of the source media starting from the first point in time when the duration of the first interpreted segment is reached.
- the media client alternates the playback between source segments consecutively followed by the recording of their respective interpreted segments.
- the interpretation of the source media starts again from the first point in time such that no fraction of the source media is left uninterpreted. In other words, the source media may be completely interpreted by the user of the media client.
- the media client is further configured to allow a user of the media client to define a mark at one or more predetermined points in time in the source media.
- the media client allows interaction with one or two or more interpreters at the same time, or with only one individual interpreted.
- a mark can comprise an audio track and/or a video track and/or text which can be then played and/or displayed by the media client for one or more interpreters.
- information can be shared using marks at one or more predetermined points in time in the source media. For example, in a preparation mode, a teacher can prepare an interpretation exercise based on an audio track or a video comprising an audio source track. Marks to indicate positions for consecutive interpreting can optionally be added to the source track and/or be made visible to the interpreter.
- marks for points of time of interest such as for example difficult words, specific vocabulary, or parts of the exercise that require special attention from the interpreters, etc.
- marks for points of time of interest can be added to the source track.
- a teacher or student interpreter can use the same application in review mode to play back the source track alongside the submitted interpreted versions.
- the teacher can make use of feedback marks to add remarks about the interpretation of the student interpreter.
- An interpreter can insert marks to a recorded interpreted segment by clicking on a specific position on the recorded interpreted segment waveform representation in the media client. In a popup window, a mark label can then be set. Together with the position, in milliseconds, or seconds, or frames, relative to the total exercise length, this information is also stored in the data, such as for example a metadata file. Alternatively, the position(s) of the marks on the recorded sequence of interpreted segments are in relation to the recorded sequence of interpreted segments only, i.e. without any relationship to the total exercise length.
- - “attention” marks are placed by a teacher in for example a preparation or a review mode onto the source waveform to indicate for example something interesting in the source media. Such a mark will for example remain visible only by a teacher;
- - “consecutive interpretation” marks are placed by a teacher in for example a preparation mode onto the source waveform to indicate the end of a sentence or to indicate the end of a part of a sentence, at the locations where the speaker is assumed to pause and the interpreter is assumed to start translate. Such a mark will be visible to for example both teacher and interpreter;
- feedback marks are placed by a teacher in for example a review mode onto the recorded waveforms to give some remark about the interpretation. Such a mark will be visible to both teacher and interpreter in review mode;
- - “student” marks are placed by a student interpreter in for example a review mode onto the recorded waveforms to give some information about the interpretation. Such a mark will be visible to both teacher and student interpreter in review mode.
- the media client is further configured to store the marks in the data.
- a teacher creates a project, for example an IQF project, based on an audio and/or video source track.
- the teacher can also set the project name, the interpreting mode (simultaneous or consecutive), the interpreting media type (audio or audio/video) and, optionally, a list of marks.
- These settings are all stored as metadata in the IQF project.
- the resulting IQF project for example comprises these settings.
- the settings are saved in an encrypted and compressed container comprising the source media and the data, such as for example a metadata file.
- the teacher can define or insert marks on the source media by clicking on a specific position on the timeline or waveform representing the source media.
- a mark type and a label can for example be set.
- this information is also stored in the data, such as for example in a metadata file.
- An interpreter typically starts by listening to a part of the source media. When the source media reaches for example a“consecutive interpretation” mark, the source media is paused.
- the recording of the microphone i.e. audio
- the webcam i.e. video
- the‘Start Recording’ button is enabled. This button becomes visible on the media client or it changes for example from an idle color to an active color. The interpreter can then click the‘Start Recording’ button. Only then the recording of the microphone, i.e. audio, and optionally the webcam, i.e. video is started.
- the interpreter When the interpreter has finished the interpretation of the first source segment, the interpreter can click the‘Stop recording’ or‘Continue playback’ button.
- the interpreter will see a‘Continue playback’ button.
- the interpreter will see a‘Stop recording’ button on the media client.
- a flat line indicating silence is being appended to the representation of the source media on the media client.
- the index and the duration of the recording of the first interpreted segment is for example added to a recordings list in the user interface of the media client.
- the media client is further configured to record again the first interpreted segment from the first point in time, thereby generating an alternative for the first interpreted segment.
- the interpreter can record multiple versions, optionally merge them, and then select a version to submit. This way, the interpreter can correct or improve his or her interpretation before submitting it for review and/or for publishing, thereby generally improving the quality of the interpretation of the source media.
- the interpreter can make as many recorded versions as necessary. Two or more recorded versions can be merged by selecting them in the versions list and clicking the‘Merge’ menu item in the media client. The recorded interpreted segments are then merged into a new recorded interpreted segments using for example ffmpeg. The metadata of the original interpreted segments is used to determine their position in the new interpreted segments. New metadata is also generated to describe the start time and durations of the slices of the resulting interpreted segments. All the recorded interpreted segments, together with the data and the source media can for example subsequently be saved into a project file, such as for example into an IQF project file.
- the media client is further configured to; - choose between the first interpreted segment and the alternative for the interpreted segment, thereby generating an updated first interpreted segment;
- the media client is further configured to:
- the media client displays in a very intuitive manner that the source media is being interpreted. Also, the media client displays in a very intuitive manner to which source segment corresponds an interpreted segment.
- the media client is further configured to display an end line after a last source segment of the source media when a duration of the sequence of interpreted segments comprising a consecutive interpretation into the second language of the source media is larger than a duration of the source media.
- the media client is further configured to simultaneously play back one or more source segments of the source media and record one or more simultaneous interpreted segments of the source segments into the second language.
- the media client allows to prepare interpreting exercises and send them to interpreters, to record performances of interpreters for simultaneous interpreting exercises, and to evaluate interpretations and add feedback for the interpreters in the recording.
- Data comprising a start time and duration of the recording of the sequence of interpreted segments is saved.
- the index and the duration of the recording can then for example be added to a recordings list.
- a resulting recorded sequence of interpreted segments consists of one audio or video file which is continuous and without interruptions.
- Information of where the sequence of interpreted segments is exactly positioned with respect to the source media can be saved in for example the data. This information comprises for example a start time in relation to the source media and a duration.
- the interpreter can insert student marks to a recorded track by clicking on a specific position on the recorded track waveform representation. In a popup window displayed on the media client, a mark label can then be set. Together with the position, in milliseconds, relative to the total exercise length, this information is also stored in the data.
- the source media plays without interruption.
- the recorded sequence of interpreted segments starts playing concurrently with the source media when the start time of the sequence of interpreted segments is met.
- the recorded sequence of interpreted segments stops playing when the position of the time mark is past the start time added with the duration of the recording sequence of interpreted segments.
- the media client is further configured to, when a duration of the one or more simultaneous interpreted segments is larger than a duration of the source segments, display an end line after the source waveform.
- the data comprises the first point in time and the second point in time or a duration of the first interpreted segment determined between the second point in time and the first point in time.
- the computer-implemented method according to the present disclosure allows to prepare interpreting exercises and send them to interpreters, to record performances of interpreters for consecutive interpreting exercises, and to evaluate interpretations and add feedback for the interpreters in the recording. Teachers or peers can evaluate these recorded interpretations rapidly and easily.
- the method according to the present disclosure evaluates not only the spoken interpretation, but also the way the interpreter presents the message. Indeed, by capturing not only for example the interpreter's microphone, but also for example the webcam of the interpreter, the performance can be properly evaluated by all means.
- the method according to the present disclosure allows a more emotional and fluent delivery of the interpretation.
- the consecutive interpreter has more time to prepare for the nuances of the language and to choose the words carefully, thereby improving the quality of the interpretation.
- the method according to the present disclosure makes it easy for interpreters to check the quality of their interpretation as well as the conformity with the source media.
- An interpreter and/or a teacher can quickly jump to points in time of interest in the source media. This is done in a user-friendly and intuitive manner.
- the method allows recording a sequence of interpreted segments of a source media. The first point in time and the second point in time are determined with respect to the source media. The steps of this method are repeated for all the interpreted segments, thereby recording a sequence of interpreted segments.
- the method further comprises the steps of:
- the source media and the sequence of interpreted segments are recorded in two separate files.
- the method according to the present disclosure is a dual-track recorder: the source media and the interpreted segments are recorded on separate tracks.
- the step of pausing the playback of the first source segment and the step of starting to record the first interpreted segment are simultaneous.
- the step of reaching the end of the first interpreted segment and the step of playing back the second source segment of the source media starting from the first point in time are simultaneous.
- the method further comprises the step of defining a mark at one or more predetermined points in time in the source media.
- the method further comprises the step of storing the marks in the data.
- the method further comprises the step of recording again the first interpreted segment from the first point in time, thereby generating a copy of the first interpreted segment.
- a media client comprising at least one processor and at least one memory including computer program code, the at least one memory and computer program code configured to, with the at least one processor, cause the media client to perform the method according to a first aspect of the present disclosure.
- a computer program product comprising computer-executable instructions for causing a media client to perform at least the method according to a first example aspect of the present disclosure.
- a computer readable storage medium comprising computer-executable instructions for performing the method according to a first example aspect of the present disclosure when the program is run on a computer.
- FIG. 1 depicts an example embodiment of a media client according to the present disclosure.
- FIG. 2 depicts an example embodiment of a source media and of a sequence of interpreted segments according to the present disclosure.
- FIG. 3 depicts an example embodiment of a consecutive interpretation with the media client according to the present disclosure.
- FIG. 4 depicts an example embodiment of a simultaneous interpretation with the media client according to the present disclosure.
- FIG. 5 depicts an example embodiment of a computer-implemented method according to the present disclosure.
- FIG. 6 shows an example embodiment of a suitable computing system for performing one or several steps in embodiments of the invention.
- FIG. 1 illustrates an example embodiment of a media client 10 according to the present disclosure for recording a sequence 20 of interpreted segments 201 ;202, wherein each interpreted segments 201 ;202 comprises a consecutive or a simultaneous interpretation into a second language of a source segment 101 ; 102 of a source media 10 comprising a first language.
- the media client 1 is configured to play back a first source segment 101 of the source media 10 comprising the first language.
- the media client is further configured to:
- the media client 1 is further configured to:
- the media client 10 pauses the playback of the first source segment 101 and starts recording the first interpreted segment 201 simultaneously.
- the media client 1 plays back the second source segment 102 of the source media 10 starting from the first point in time 91 when the duration 7 of the first interpreted segment 201 is reached.
- the media client 1 allows a user of the media client 1 to define a mark 50 at one or more predetermined points in time 80 in the source media 10.
- the media client 1 can optionally store the marks 50 in the data 30.
- the media client 1 can record again the first interpreted segment 201 from the first point in time 91 , thereby generating a copy of the first interpreted segment.
- the media client 1 can:
- the media client 1 is further configured to:
- the media client 1 is further configured to display an end line after a last source segment of said source media 10 when a duration of the sequence of interpreted segments comprising a consecutive interpretation into the second language of the source media 10 is larger than a duration of the source media 10.
- the media client 1 is further configured to simultaneously play back the source media 10 and record a simultaneous interpreted segment of the source media 10 into the second language, and when a duration of the simultaneous interpreted segment is larger than a duration of the source media 10, the media client 10 is further configured to display an end line 63 after the source waveform 60.
- the media client 1 is further configured to simultaneously play back the source media 10 and record a simultaneous live interpreted segment 201 ;202 of the source media 10 into the second language.
- FIG. 2 illustrates an example embodiment of a sequence 20 of interpreted segments 201 ;202, wherein each interpreted segments 201 ;202 comprises a consecutive or a simultaneous interpretation into a second language of a source segment 101 ; 102 of a source media 10 comprising a first language.
- Data 30 on the sequence of interpreted segments 201 ;202 is stored in for example the media client of Fig.
- the data 30 comprises one or more start points in time 9 of interpreted segments with respect to the source media 10, such as for example the first point in time 91 and the second point in time 92, and further comprises a duration 6 of the interpreted segments, such as the duration 7 of the first interpreted segment 201 calculated between the second point in time 92 and the first point in time 91 and the duration 8 of the second interpreted segment 202 calculated between a fourth point in time 94 and a third point in time 93.
- FIG. 3 illustrates an example embodiment of a consecutive interpretation by the media client 1 of Fig. 1.
- a source media 10 comprises one or more source segments 101 ; 102.
- the first source segment starts in time at T1 .
- the media client then plays back the first source segment 101 comprising a first language from the moment 90 to the first point in time 91.
- the media client pauses the play back of the first source segment 101 and records a live interpretation of the source segment 101 by a user of the media client.
- the media client records a first interpreted segment 201 comprising a consecutive interpretation into a second language different from the first language of the first source segment 101.
- the media client pauses the recording of the first interpreted segment 201 , for example when the user of the media client is done with interpreting the first source segment 101 , and the media client plays back a second source segment 102 of the source media 10 starting again from the first point in time 91.
- the media client determines a duration 7 of the first interpreted segment 201 by calculating a time difference between the second point in time 92 and the first point in time 91 , and saves this duration 7 as a duration 6 in data 30.
- the media client then stores data 30, for example metadata related to the first interpreted segment 201 , wherein the metadata 30 comprises a start time 9 of the first interpreted segment 201 with respect to the source media 10 and a duration 7 of the first interpreted segment 201.
- the start time 9 is 91 and the duration 7 is equal to 92-91.
- the media client pauses the play back of the second source segment 102 and records a second interpreted segment 202 comprising a consecutive interpretation into another language of the second source segment 102. This other language is different from the first language, but it can be identical or different to the second language.
- the media client pauses the recording of the second interpreted segment 202, for example when the user of the media client is done with interpreting the second source segment 102, and the media client plays back a third source segment of the source media 10 starting again from the other point in time 93.
- the media client determines a duration 8 of the second interpreted segment 202 by calculating a time difference between the point in time 94 and the other point in time 93, and saves this duration 8 as a duration 6 in data 30.
- the media client then stores data 30, for example metadata related to the second interpreted segment 202, wherein the metadata 30 comprises a start time 9 of the second interpreted segment 202 with respect to the source media 10 and a duration 8 of the second interpreted segment 202.
- the start time 9 is 93 and the duration 8 is equal to 94-93.
- the media client of Fig. 1 repeats this method for all the interpreted segments, thereby recording a sequence of interpreted segments 20, and being able to play back an alternating media comprising source segments 101 ; 102 and interpreted segments 201 ;202 alternating during playback.
- the source segments comprise a first language and the interpreted segments comprise a second language different from the first language.
- FIG. 4 illustrates an example embodiment of a simultaneous interpretation by the media client 1 of Fig. 1.
- a source media 10 comprises one source segment 101.
- the media client then plays back the source segment 101 of the media source 10 comprising a first language from the moment 90 to the first point in time 91.
- the media client records a live interpretation of the source segment 101 by a user of the media client.
- the media client records a first interpreted segment 201 comprising a consecutive interpretation into a second language different from the first language of the first source segment 101.
- the source media 10 is over and the media client continues the recording of a live interpretation 201 of the source segment 101 by a user of the media client until the second point in time 92.
- the media client determines a duration 7 of the first interpreted segment 201 by calculating a time difference between the first point in time 91 and the moment 90, and saves this duration 7 as a duration 6 in data 30.
- the media client then stores data 30, for example metadata related to the first interpreted segment 201 , wherein the metadata 30 comprises a start time 9 of the first interpreted segment 201 with respect to the source media 10 and a duration 7 of the first interpreted segment 201.
- the media client is further configured to simultaneously play back the source media 10 and record a simultaneous interpreted segment of the source media 10 into the second language, and when a duration of the simultaneous interpreted segment is larger than a duration of the source media 10, the media client 10 is further configured to display an end line 64 after the source waveform 60.
- FIG. 5 schematically illustrates the steps of a method according to an example embodiment for recording a sequence of interpreted segments 201 ;202, wherein each interpreted segment 201 ;202 comprises a consecutive interpretation into a second language of a source segment 101 ; 102 of a source media 10 comprising a first language.
- a first step 501 a first source segment 101 of the source media 10 comprising the first language is played back.
- a second step 502 consecutive to the first step 501 , and at a first point in time 91 , the playback of the first source segment 101 is paused and a first interpreted segment 201 comprising a consecutive interpretation into the second language of the first source segment 101 is recorded.
- a third step 503 consecutive to the second step 502, and at a second point in time 92, the recording of the first interpreted segment 201 is paused and a second source segment 102 of the source media 10 starting from the first point in time 91 is played back.
- a fourth step 504 consecutive to the third step 503, a duration of the first interpreted segment 201 is determined between the second point in time 92 and the first point in time 91.
- data 30 on the first interpreted segment 201 is stored, wherein the data 30 comprises the first point in time 91 and the duration 7 of the first interpreted segment 201.
- FIG. 6 shows a suitable computing system 800 enabling to implement embodiments of the system.
- Computing system 800 may in general be formed as a suitable general-purpose computer and comprise a bus 810, a processor 802, a local memory 804, one or more optional input interfaces 814, one or more optional output interfaces 816, a communication interface 812, a storage element interface 806, and one or more storage elements 808.
- Bus 810 may comprise one or more conductors that permit communication among the components of the computing system 800.
- Processor 802 may include any type of conventional processor or microprocessor that interprets and executes programming instructions.
- Local memory 804 may include a random-access memory (RAM) or another type of dynamic storage device that stores information and instructions for execution by processor 802 and/or a read only memory (ROM) or another type of static storage device that stores static information and instructions for use by processor 802.
- Input interface 814 may comprise one or more conventional mechanisms that permit an operator or user to input information to the computing device 800, such as a keyboard 820, a mouse 830, a pen, voice recognition and/or biometric mechanisms, a camera, etc.
- Output interface 816 may comprise one or more conventional mechanisms that output information to the operator or user, such as a display 840, etc.
- Communication interface 812 may comprise any transceiver-like mechanism such as for example one or more Ethernet interfaces that enables computing system 800 to communicate with other devices and/or systems, for example with other computing devices 881 , 882, 883.
- the communication interface 812 of computing system 800 may be connected to such another computing system by means of a local area network (LAN) or a wide area network (WAN) such as for example the internet.
- Storage element interface 806 may comprise a storage interface such as for example a Serial Advanced Technology Attachment (SATA) interface or a Small Computer System Interface (SCSI) for connecting bus 810 to one or more storage elements 808, such as one or more local disks, for example SATA disk drives, and control the reading and writing of data to and/or from these storage elements 808.
- SATA Serial Advanced Technology Attachment
- SCSI Small Computer System Interface
- computing system 800 could thus correspond to the media client 1 in an embodiment of the present disclosure for example illustrated by Fig. 1 .
- circuitry may refer to one or more or all of the following:
- circuitry also covers an implementation of merely a hardware circuit or processor (or multiple processors) or portion of a hardware circuit or processor and its (or their) accompanying software and/or firmware.
- circuitry also covers, for example and if applicable to the particular claim element, a baseband integrated circuit or processor integrated circuit for a mobile device ora similar integrated circuit in a server, a cellular network device, or other computing or network device.
- top, bottom, over, under, and the like are introduced for descriptive purposes and not necessarily to denote relative positions. It is to be understood that the terms so used are interchangeable under appropriate circumstances and embodiments of the invention are capable of operating according to the present invention in other sequences, or in orientations different from the one(s) described or illustrated above.
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Physics & Mathematics (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201962879664P | 2019-07-29 | 2019-07-29 | |
PCT/EP2020/069661 WO2021018555A1 (en) | 2019-07-29 | 2020-07-10 | Media client for recording and playing back interpretation |
Publications (1)
Publication Number | Publication Date |
---|---|
EP4004918A1 true EP4004918A1 (en) | 2022-06-01 |
Family
ID=71579624
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20739675.5A Pending EP4004918A1 (en) | 2019-07-29 | 2020-07-10 | Media client for recording and playing back interpretation |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP4004918A1 (en) |
AU (1) | AU2020321316A1 (en) |
WO (1) | WO2021018555A1 (en) |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100399999B1 (en) * | 2001-02-05 | 2003-09-29 | 삼성전자주식회사 | Recording medium containing multi-stream recorded thereon, recording apparatus, recording method therefor, reproducing apparatus, and reproducing method therefor |
US20070106516A1 (en) * | 2005-11-10 | 2007-05-10 | International Business Machines Corporation | Creating alternative audio via closed caption data |
LU91549A2 (en) * | 2009-04-07 | 2010-10-07 | European Community | Online speech repository and client tool therefor |
US8744239B2 (en) * | 2010-08-06 | 2014-06-03 | Apple Inc. | Teleprompter tool for voice-over tool |
-
2020
- 2020-07-10 EP EP20739675.5A patent/EP4004918A1/en active Pending
- 2020-07-10 AU AU2020321316A patent/AU2020321316A1/en active Pending
- 2020-07-10 WO PCT/EP2020/069661 patent/WO2021018555A1/en unknown
Also Published As
Publication number | Publication date |
---|---|
WO2021018555A1 (en) | 2021-02-04 |
AU2020321316A1 (en) | 2022-02-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Evers | From the past into the future. How technological developments change our ways of data collection, transcription and analysis | |
Burston | Video dubbing projects in the foreign language curriculum | |
US8113844B2 (en) | Method, system, and computer-readable recording medium for synchronous multi-media recording and playback with end user control of time, data, and event visualization for playback control over a network | |
Rohlfing et al. | Comparison of multimodal annotation tools-workshop report | |
WO2018227761A1 (en) | Correction device for recorded and broadcasted data for teaching | |
Pavel et al. | VidCrit: video-based asynchronous video review | |
US10553129B2 (en) | System and method for recording, documenting and visualizing group conversations | |
Wang et al. | Voicecoach: Interactive evidence-based training for voice modulation skills in public speaking | |
Thompson | Building a specialised audiovisual corpus | |
Tuncer et al. | Making time: Pausing to coordinate video instructions and practical tasks | |
US20150213793A1 (en) | Methods and systems for converting text to video | |
Alksne | How to produce video lectures to engage students and deliver the maximum amount of information | |
Nuari et al. | Using camtasia studio 8 to produce learning video to teach english through e-learning | |
McIlvenny | New technology and tools to enhance collaborative video analysis in live ‘data sessions’ | |
Lewis et al. | Musicological observations during rehearsal and performance: A Linked Data digital library for annotations | |
Bakkay et al. | Protocols and software for simplified educational video capture and editing | |
Chi et al. | Synthesis-Assisted Video Prototyping From a Document | |
Magalhães | Music, performance, and preservation: insights into documentation strategies for music theatre works | |
EP4004918A1 (en) | Media client for recording and playing back interpretation | |
US20080222505A1 (en) | Method of capturing a presentation and creating a multimedia file | |
Baume et al. | A contextual study of semantic speech editing in radio production | |
Baume | Semantic Audio Tools for Radio Production | |
Loehr et al. | Commonplace tools for studying commonplace interactions: practitioners' notes on entry-level video analysis | |
Minutella | Audio description software tools | |
Mora | Creation of educational videos: tools and tips |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20220228 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |