US20140006020A1 - Transcription method, apparatus and computer program product - Google Patents
Transcription method, apparatus and computer program product Download PDFInfo
- Publication number
- US20140006020A1 US20140006020A1 US13/538,671 US201213538671A US2014006020A1 US 20140006020 A1 US20140006020 A1 US 20140006020A1 US 201213538671 A US201213538671 A US 201213538671A US 2014006020 A1 US2014006020 A1 US 2014006020A1
- Authority
- US
- United States
- Prior art keywords
- audio file
- text report
- transcripted text
- transcripted
- words
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013518 transcription Methods 0.000 title claims abstract description 37
- 230000035897 transcription Effects 0.000 title claims abstract description 37
- 238000004590 computer program Methods 0.000 title claims abstract description 35
- 238000000034 method Methods 0.000 title claims abstract description 33
- 238000012545 processing Methods 0.000 claims description 54
- 230000003190 augmentative effect Effects 0.000 claims description 4
- 230000000875 corresponding effect Effects 0.000 description 43
- 230000015654 memory Effects 0.000 description 24
- 230000006870 function Effects 0.000 description 18
- 238000012552 review Methods 0.000 description 18
- 238000004891 communication Methods 0.000 description 7
- 230000002596 correlated effect Effects 0.000 description 7
- 230000003416 augmentation Effects 0.000 description 6
- 230000008569 process Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 238000002591 computed tomography Methods 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 238000009607 mammography Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000002604 ultrasonography Methods 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000002600 positron emission tomography Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/61—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Definitions
- Embodiments of the present invention relate generally to the transcription of an audio file and, more particularly, to the augmentation of a transcripted text report of an audio file to facilitate, for example, subsequent review of the transcripted text report in light of the associated audio file.
- transcripted text report It is common in a number of scenarios to dictate a report or the like which is captured in an audio file.
- An audio file is oftentimes transcribed in order to create a transcripted text report.
- the transcripted text report may then be saved as a written record and may be referenced by others in order to review the report.
- a radiologist may review an image of a patient, such as an x-ray or other image, and may dictate a report relating to the image including the radiologist's findings with respect to the image. Following dictation of the report, the report may be typed by a transcriptionist to create the corresponding transcripted text report. A healthcare practitioner, such as the same or another radiologist, will then review the transcripted text report to determine if the transcripted text report is accurate. In an instance in which the radiologist who reviews the transcripted text report determines that the transcripted text report is accurate, the reviewing radiologist may sign off or otherwise verify the transcripted text report.
- the inaccuracy may be flagged or otherwise identified for correction.
- the reviewing radiologist may listen to the entire audio file. The reviewing radiologist may then wish to listen again to selected portions of the audio file, such as the portions of the audio file that present the findings or other final impressions. In order to reposition the audio file in order to listen to the selected portions of the audio file, the reviewing radiologist may need to fast forward or rewind the audio file and may therefore be required to estimate the proper position within the audio file at which the selected portion of the audio file is located.
- This process of reviewing the audio file in conjunction with the determination of the accuracy of a transcripted text report may sometimes be inefficient and time consuming since the reviewing radiologist may be required to repeatedly guess as to the portion of the audio file to be reviewed with the radiologist's guesses not always proving to be correct. Such inefficiencies may be particularly challenging for healthcare practitioner, such as radiologist, who already has a demanding schedule.
- a transcription method, apparatus and computer program product are provided according to embodiments of the present invention in order to permit a transcripted text report to be reviewed in a more efficient manner.
- the transcription method, apparatus and computer program product of an example embodiment may correlate portions of the transcripted text report and the audio file in such a manner that corresponding portions of the transcripted text report and the audio file may be readily identified, thereby facilitating the playback of the a selected portion of the audio file during the review of the transcripted text report.
- a transcripted text report may be reviewed for accuracy relative to a corresponding audio file in a quicker and more efficient manner.
- a method in one embodiment, includes receiving an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the method of this embodiment correlates, with processing circuitry, the respective position within the transcripted text report with a corresponding position within the audio file. Although various positions within the transcripted text report may be identified, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or the commencement of a paragraph. The method of this embodiment also augments the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with a respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the plurality of selectable elements may include a plurality of icons.
- the augmentation of the transcripted text report may include the insertion of an icon within the transcripted text report at each of the plurality of positions of the transcripted text report.
- the icons may be responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the icons of one embodiment are also responsive to user actuation in order to cause the audio file to be played beginning at a corresponding position.
- the icons of one embodiment are also responsive to user actuation in order to cause the playing of the audio file to be paused.
- the correlation of the respective position within the transcripted text report with a corresponding position within the audio file may include utilizing voice recognition to analyze the audio files so as to identify words and comparing the words that are identified from the audio file with words of the transcripted text report to identify a correspondence there between.
- the comparison of the words that are identified from the audio file with words of the transcripted report may include the comparison of a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify a correspondence therebetween.
- a transcription apparatus in another embodiment, includes processing circuitry configured to receive an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the processing circuitry is configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file. Although the plurality of positions within the transcripted text report may be determined in various manners, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or the commencement of a paragraph.
- the processing circuitry of this embodiment is also configured to augment the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with a respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the plurality of selectable elements of one embodiment includes a plurality of icons.
- the processing circuitry is configured to augment the transcripted text report by inserting an icon within the transcripted text report at each of the plurality of positions of the transcripted text report.
- the icons of this embodiment are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the icons may also be responsive to user actuation in order to cause the audio file to be played beginning at the corresponding position. In an instance in which the audio file is already being played, the icons may alternatively be responsive to user actuation in order to cause playing of the audio file to be paused.
- the processing circuitry of one embodiment is configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file by utilizing voice recognition to analyze the audio file so as to identify words and then to comparing the words that are identified from the audio file with words of the transcripted text report to identify a correspondence therebetween.
- the processing circuitry may be configured to compare the words that are identified from the audio file with words from the transcripted text report by comparing a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify a correspondence therebetween.
- a computer program product includes at least one non-transitory computer-readable storage medium having computer-executable program code instructions stored therein.
- the computer-executable program code instructions include program instructions configured to receive an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the program code instructions of this embodiment are configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file. While a plurality of positions may be determined in various manners, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or commencement of a paragraph.
- the computer-executable program code instructions of this embodiment also include program code instructions configured to augment the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with the respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the plurality of selectable elements of one embodiment includes a plurality of icons.
- the program code instructions configured to augment the transcripted text report include program code instructions configured to insert an icon within the transcripted text report at each of a plurality of positions of the transcripted text report.
- the icons may be responsive to user actuation in order to cause the audio file to move to the corresponding position.
- the icons may also be responsive to user actuation in order to cause the audio file to be played beginning at the corresponding position or to cause playing of the audio file to be paused.
- the program code instructions configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file may, in one embodiment, include program code instructions configured to utilize voice recognition to analyze the audio file to identify words and program code instructions configured to compare the words that are identified from the audio file with words of the transcripted text report to identify correspondence therebetween.
- the program code instructions that are configured to compare the words that are identified from the audio file with words of a transcripted text report include, in one embodiment, program code instructions configured to compare a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify correspondence therebetween.
- FIG. 1 is a flow chart illustrating the operations performed by a transcription apparatus configured in accordance with an example embodiment of the present invention
- FIG. 2 is a block diagram of an transcription apparatus that may be specifically configured in accordance with an example embodiment of the present invention
- FIG. 3 is an example of a transcripted text report following augmentation in accordance with an example embodiment of the present invention
- FIG. 4 is an example of the transcripted text report of FIG. 3 following augmentation in accordance with an example embodiment of the present invention and following selection of an icon within the transcripted text report;
- FIG. 5 is an example of the transcripted text report of FIG. 3 following augmentation in accordance with an example embodiment of the present invention and following selection of another icon within the transcripted text report.
- a transcription method, apparatus and compute program product are provided in accordance with example embodiments of the present invention in order to correlate an audio file and a transcripted text report of the audio file.
- the transcription method, apparatus and computer program product of an example embodiment may facilitate the review and verification of the transcripted text report by reference to the corresponding audio file.
- the transcripted text report may be reviewed and verified in a more efficient and a quicker manner since those portions of the audio file that correspond to portions of interest within the transcripted text report may be readily identified and reviewed.
- transcription method, apparatus and computer program product may be utilized in conjunction with any type of audio file and any type of transcripted text report
- the transcription method, apparatus and computer program product of an example embodiment will be described hereinafter for purposes of example, but not of limitation, in conjunction with the review of an audio file created by a radiologist, such as during or following the review of an image by a radiologist, in order to generate a report that includes the radiologist's findings with respect to the image and that has been verified relative to the audio file.
- an audio file may be initially created, such as by dictating to a digital voice recorder or other type of dictation device.
- the audio file may relate to various subjects including, for example, the reports generated by health care professionals.
- a radiologist may review an image of a patient and may dictate a report relating to the review of the image, including their findings with respect to the image.
- the image may be acquired by a number of different imaging modalities including, for example, ultrasound (US), magnetic resonance (MR), positron emission tomography (PET), computed tomography (CT), mammography (MG), digital radiology (DR), computed radiology (CR) or the like.
- the audio file may then be transcribed, such as by a transcriptionist, to create a transcripted text file.
- the transcripted text file is reviewed and verified relative to the corresponding audio file.
- a transcripted text report corresponding to an audio file that represents the report of a health care practitioner, such as report of a radiologist following the review of an image may be performed to verify that the transcripted text file is an accurate reproduction of the audio file.
- a radiologist such as a radiologist who initially dictated the audio file or another radiologist, may review the transcripted text file to verify its accuracy, including listening to all or at least portions of the audio file in order to make such a determination.
- the method of one embodiment may receive an audio file and a transcripted text report corresponding to the audio file, that is, the transcripted text report that represents a transcription of the audio file. See block 10 of FIG. 1 .
- the operations represented by the flow chart of FIG. 1 may be performed by a transcription apparatus 20 that is specifically configured in accordance with embodiments of the present invention.
- a transcription apparatus may be embodied in any of a variety of computing devices, such as laptop computers, tablet computers, smartphones, desktop computers, computer workstations or the like.
- the transcription apparatus includes various means for performing the various functions described herein.
- the means of the apparatus as described herein may be embodied as, for example, circuitry, hardware elements (e.g., a suitably programmed processor, combinational logic circuit, and/or the like), a computer program product comprising computer-readable program instructions (e.g., software or firmware) stored on a computer-readable medium (e.g. memory) that is executable by a suitably configured processing device (e.g., the processing device), or some combination thereof.
- a suitably programmed processor e.g., a suitably programmed processor, combinational logic circuit, and/or the like
- a computer program product comprising computer-readable program instructions (e.g., software or firmware) stored on a computer-readable medium (e.g. memory) that is executable by a suitably configured processing device (e.g., the processing device), or some combination thereof.
- the processing circuitry 22 may, for example, be embodied as various means including one or more microprocessors, one or more coprocessors, one or more multi-core processors, one or more controllers, one or more computers, various other processing elements including integrated circuits such as, for example, an ASIC (application specific integrated circuit) or FPGA (field programmable gate array), or some combination thereof.
- the processing circuitry may be embodied on a single computing device or may be distributed across a plurality of computing devices collectively configured to function as the apparatus.
- the processing circuitry is configured to execute instructions stored in the memory 24 or otherwise accessible to the processing circuitry. These instructions, when executed by the processing circuitry, may cause the transcription apparatus 20 to perform one or more of the functionalities of the apparatus as described herein.
- the processing circuitry may comprise an entity capable of performing operations according to embodiments of the present invention while configured accordingly.
- the processing circuitry when the processing circuitry is embodied as an ASIC, FPGA or the like, the processing circuitry may comprise specifically configured hardware for conducting one or more operations described herein.
- the processing circuitry when the processing circuitry is embodied as an executor of instructions, such as may be stored in the memory, the instructions may specifically configure the processing circuitry to perform one or more algorithms and operations described herein.
- the memory 24 may include, for example, volatile and/or non-volatile memory. Although illustrated in FIG. 2 as a single memory, the memory may comprise a plurality of memories. The plurality of memories may be embodied on a single computing device or distributed across a plurality of computing devices.
- the memory may comprise, for example, a hard disk, random access memory, cache memory, flash memory, an optical disc (e.g., a compact disc read only memory (CD-ROM), digital versatile disc read only memory (DVD-ROM), or the like), circuitry configured to store information, or some combination thereof.
- the memory may comprise any non-transitory computer readable storage medium.
- the memory may be configured to store information, data, applications, instructions, or the like for enabling the transcription apparatus 20 to carry out various functions in accordance with example embodiments of the present invention.
- the memory is configured to buffer input data for processing by the processing circuitry 22 .
- the memory is configured to store program instructions for execution by the processing circuitry.
- the user interface 26 may be in communication with the processing circuitry 22 to receive an indication of a user input and/or to provide an audible, visual, mechanical, or other output to a user.
- the user interface may include, for example, a keyboard and/or a touch sensitive input device, such as a touch sensitive mouse, for receiving user input.
- the user interface may also include a display for presenting medical images for review and manipulation by a user.
- the user interface may be in communication with the memory 24 , such as via a bus.
- the transcription apparatus 20 of one embodiment may also include a communication interface, such as one or more antennas, one or more transceivers, etc., for facilitating communication between the transcription apparatus and one or more other computing or communications devices.
- the transcription apparatus may include means, such as the processing circuitry 22 , the communication interface or the like, for receiving the audio file and the transcripted text report that corresponds to the audio file.
- the audio file and the transcripted text report may be stored in memory 24 and accessed by the processing circuitry.
- the transcription apparatus may include a communication interface for receiving the audio file and/or the transcription text report, such as via a wireless connection or a wireline connection.
- the transcription apparatus 20 may include means, such as the processing circuitry 22 or the like, for correlating, for each of a plurality of positions within the transcripted text report, a respective position within the transcripted text report with a corresponding position within the audio file. For example, each word of the transcripted text report may be correlated with a corresponding position within the audio file, each sentence within the transcripted text report may be associated with a corresponding position within the audio file, and/or each paragraph within the transcripted text report may be correlated with a corresponding position within the audio file.
- the various positions within the transcripted text report that are correlated with corresponding positions within the audio file may be identified in other manners, such as by correlating each punctuation mark within the transcripted text report with a corresponding position within the audio file.
- the correlation between a respective position within the transcripted text report and a corresponding position within the audio file may be performed by the processing circuitry 22 in various manners.
- the processing circuitry utilizes voice recognition to analyze the audio file so as to identify words.
- the processing circuitry may implement a voice recognition engine that is configured to recognize and to identify the pronunciation of a plurality of words and, in one embodiment, that may be trained so as to recognize the pronunciation of a plurality of words, such as a plurality of the most frequently used words, by a particular user, such as a radiologist.
- the processing circuitry 22 and, more particularly, the voice recognition engine may identify each unique word that is included within the audio file. For example, if the audio file includes a following sentence, “I am working on it, and I am close.”
- the processing circuitry such as a voice recognition engine of the processing circuitry, may identify “I”, “am”, “working”, “on”, “it” and “close” as the unique words that are included within the audio file.
- the processing circuitry and, more particularly, the voice recognition engine implemented by the processing circuitry also identifies the position of each word within the audio file. In this regard, the position of a word within the audio file will be hereinafter referenced as the audio position of the word.
- the processing circuitry and, more particularly, the voice recognition engine of one embodiment may identify the audio position of each word based upon the elapsed time from the beginning of the audio file to the word as the audio position of the word in one embodiment.
- the processing circuitry 22 and, more particularly, the voice recognition engine implemented by the processing circuitry may identify both audio positions of the word “I”, such as at one millisecond and five milliseconds relative to the beginning of the audio file.
- the processing circuitry such as a voice recognition engine implemented by the processing circuitry, may identify the position of each occurrence of the word within the audio file.
- the position of a respective word that appears multiple times may be represented as a compact form, such as [“I” 1 ms, 5 ms], the compact representation may be expanded such that each occurrence of the same word is individually associated with a corresponding position within the audio file, such as [“I” 1 ms] and [“I” 5 ms].
- the association of a word and its corresponding audio location may be referred to as a word and audio location pair, with a plurality of word and audio location pairs, such as all of the word and audio location pairs that appear within the audio file, being referred to as a collection. Based upon the respective audio locations, the collection of word and audio location pairs may be placed in an order of occurrence.
- the processing circuitry 22 may compare subsequences from the collection of word and audio location pairs that have been identified from the audio file with sentences from the transcripted text report on a sentence-by-sentence basis in order to determine a correspondence therebetween. For each sentence of the transcripted text report, the processing circuitry, such as the voice recognition engine implemented by the processing circuitry of one embodiment, may identify the longest subsequence within the collection of words and audio location pairs that matches the respective sentence of the transcripted text report and may identify the resulting subsequence of word and audio location pairs to correspond to the respective sentence.
- the transcription apparatus 20 and, more particularly, the processing circuitry 22 , the memory 24 or the like may also augment the transcripted text report to include a plurality of selectable elements. See block 14 of FIG. 1 .
- each selectable element is associated with a respective position within the transcripted text report.
- the processing circuitry of one embodiment may be configured to associate a selectable element with the beginning of each paragraph and/or with each punctuation mark within the transcripted text report.
- the selectable element or at least the respective position within the transcripted text report with which the selectable element is associated may also be associated with a corresponding position within the audio file based upon the correlation of the respective position within the transcripted text report and the corresponding position within the audio file.
- each selectable element that is associated with the beginning of a paragraph from the transcripted text report may also be associated with the corresponding position within the audio file that is correlated to the beginning of the respective paragraph.
- each selectable element that is associated with a punctuation mark within the transcripted text report may also be associated with the corresponding position within the audio file that is correlated to the respective punctuation mark.
- a user who is reviewing the transcripted text report may actuate a selectable element in order to cause the audio file to move to the corresponding position, that is, to move to the position within the audio file that has been correlated with the position within the transcripted text report with which the selectable element is associated.
- a transcripted text report may be augmented with a plurality of icons that serve as the selectable elements.
- the transcripted text report may be augmented by inserting an icon within the transcripted text report at each of the plurality of positions of the transcripted text report that are to be associated with selectable elements.
- FIG. 3 illustrates an initial portion of a speech by President Obama that is entitled “Yes We Can”.
- an icon has been associated with the beginning of each paragraph as well as with each punctuation mark.
- the transcription method and apparatus of one embodiment will cause the audio file to be moved to the corresponding position that has been correlated with the position within the transcripted text report that is associated with the icon that has been selected.
- selection of the icon designated 30 a in FIG. 3 will cause the audio file to move to a position designated 1:22 as shown by the time scroll bar 32 of FIG. 4 .
- selection of the icon designated 30 b in FIG. 3 will cause the audio file to move to the position designated 1:54 as shown by the time scroll bar 32 of in FIG. 5 .
- the inclusion of icons or other selectable elements within the transcripted text report permits a user to quickly and accurately move to different portions of the audio file in order to listen to the portion of the audio file that corresponds to the portion of the transcripted text report associated with the icon that has been selected.
- the selection of an icon not only causes the audio file to move to the position associated with the selected icon, but also to commence playing of the audio file at the selected position.
- the user may then click any of the icons as shown in FIGS. 4 and 5 .
- the icons may include a representation of the function, e.g., play, pause, etc., associated with the icon.
- the icons 30 may include a representation of a “play” button as shown in FIG. 3 so as to provide a visual indication to the user regarding the associated function that is initiated upon selection of the icon.
- the representation of the icons 34 may be changed as shown in FIGS. 4 and 5 to include a representation of a “pause” button so as to provide a visual indication to the user regarding the function that would be initiated upon actuation of the icon.
- the selection of an icon in an instance in which the audio file is being played and the selection of an icon at the same position within the transcripted text report in an instance in which the audio file is not being played may initiate different functions, such as with the function brought about by actuation of the icon alternately toggling between playing of the selected portion of the audio file in an instance in which the audio file is not currently being played and ceasing further playing of the audio file in an instance in which the audio file is currently being played.
- the icons 34 may also serve as a mechanism to permit the volume to be readily adjusted.
- a user may place the cursor in alignment with a respective icon and may then provide an input, such as rotation of a scroll wheel carried by a computer mouse, in order to controllably adjust the volume of the audio file being played.
- a user may efficiently navigate through the transcripted text file and may review the corresponding portions of the audio file without having to guess as to the relevant portions of the audio file.
- the radiologist may review the transcripted text file in conjunction with the audio file and may quickly and efficiently replay any selected portions of the audio file, such as those portions of the audio file that capture the key findings of the radiologist.
- the transcription method, apparatus and computer program product may improve the overall efficiency of the report dictation, transcription and verification process.
- FIG. 1 illustrates a flowchart of a system, method, and computer program product according to example embodiments of the invention. It will be understood that each block of the flowchart, and combinations of blocks in the flowchart, may be implemented by various means, such as hardware and/or a computer program product comprising one or more computer-readable mediums having computer readable program instructions stored thereon. For example, one or more of the procedures described herein may be embodied by computer program instructions of a computer program product. In this regard, the computer program product(s) which embody the procedures described herein may be stored by one or more memory devices 24 of a computing device and executed by processing circuitry 22 .
- the computer program instructions comprising the computer program product(s) which embody the procedures described above may be stored by memory devices of a plurality of computing devices.
- any such computer program product may be loaded onto a computer or other programmable apparatus to produce a machine, such that the computer program product including the instructions which execute on the computer or other programmable apparatus creates means for implementing the functions specified in the flowchart block(s).
- the computer program product may comprise one or more computer-readable memories on which the computer program instructions may be stored such that the one or more computer-readable memories can direct a computer or other programmable apparatus to function in a particular manner, such that the computer program product comprises an article of manufacture which implements the function specified in the flowchart block(s).
- the computer program instructions of one or more computer program products may also be loaded onto a computer or other programmable apparatus to cause a series of operations to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions which execute on the computer or other programmable apparatus implement the functions specified in the flowchart block(s).
- blocks or steps of the flowchart support combinations of means for performing the specified functions and combinations of steps for performing the specified functions. It will also be understood that one or more blocks of the flowchart, and combinations of blocks in the flowchart, may be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or combinations of special purpose hardware and computer program product(s).
- any suitable means for carrying out each of the functions described above may be employed to carry out embodiments of the invention.
- suitably configured processing circuitry 22 may provide all or a portion of the elements of the invention.
- all or a portion of the elements of the invention may be configured by and operate under control of a computer program product.
- the computer program product for performing the methods of embodiments of the invention includes a computer-readable storage medium, such as the non-volatile storage medium, and computer-readable program code portions, such as a series of computer instructions, embodied in the computer-readable storage medium.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Document Processing Apparatus (AREA)
Abstract
Description
- Embodiments of the present invention relate generally to the transcription of an audio file and, more particularly, to the augmentation of a transcripted text report of an audio file to facilitate, for example, subsequent review of the transcripted text report in light of the associated audio file.
- It is common in a number of scenarios to dictate a report or the like which is captured in an audio file. An audio file is oftentimes transcribed in order to create a transcripted text report. The transcripted text report may then be saved as a written record and may be referenced by others in order to review the report.
- By way of example, a radiologist may review an image of a patient, such as an x-ray or other image, and may dictate a report relating to the image including the radiologist's findings with respect to the image. Following dictation of the report, the report may be typed by a transcriptionist to create the corresponding transcripted text report. A healthcare practitioner, such as the same or another radiologist, will then review the transcripted text report to determine if the transcripted text report is accurate. In an instance in which the radiologist who reviews the transcripted text report determines that the transcripted text report is accurate, the reviewing radiologist may sign off or otherwise verify the transcripted text report. Alternatively, in an instance in which the transcripted text report contains an inaccuracy, the inaccuracy may be flagged or otherwise identified for correction. By reviewing the transcripted text report and verifying the accuracy of the transcripted text report, it is ensured that the transcriptionist has correctly transcribed the audio file including, most importantly, the radiologist's findings or other final impressions of the image.
- In order to review the accuracy of the transcripted text report, the reviewing radiologist may listen to the entire audio file. The reviewing radiologist may then wish to listen again to selected portions of the audio file, such as the portions of the audio file that present the findings or other final impressions. In order to reposition the audio file in order to listen to the selected portions of the audio file, the reviewing radiologist may need to fast forward or rewind the audio file and may therefore be required to estimate the proper position within the audio file at which the selected portion of the audio file is located. This process of reviewing the audio file in conjunction with the determination of the accuracy of a transcripted text report may sometimes be inefficient and time consuming since the reviewing radiologist may be required to repeatedly guess as to the portion of the audio file to be reviewed with the radiologist's guesses not always proving to be correct. Such inefficiencies may be particularly challenging for healthcare practitioner, such as radiologist, who already has a demanding schedule.
- A transcription method, apparatus and computer program product are provided according to embodiments of the present invention in order to permit a transcripted text report to be reviewed in a more efficient manner. In this regard, the transcription method, apparatus and computer program product of an example embodiment may correlate portions of the transcripted text report and the audio file in such a manner that corresponding portions of the transcripted text report and the audio file may be readily identified, thereby facilitating the playback of the a selected portion of the audio file during the review of the transcripted text report. Thus, a transcripted text report may be reviewed for accuracy relative to a corresponding audio file in a quicker and more efficient manner.
- In one embodiment, a method is provided that includes receiving an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the method of this embodiment correlates, with processing circuitry, the respective position within the transcripted text report with a corresponding position within the audio file. Although various positions within the transcripted text report may be identified, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or the commencement of a paragraph. The method of this embodiment also augments the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with a respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- The plurality of selectable elements may include a plurality of icons. In this embodiment, the augmentation of the transcripted text report may include the insertion of an icon within the transcripted text report at each of the plurality of positions of the transcripted text report. The icons may be responsive to user actuation in order to cause the audio file to move to the corresponding position. The icons of one embodiment are also responsive to user actuation in order to cause the audio file to be played beginning at a corresponding position. Alternatively, in an instance in which the audio file is already being played, the icons of one embodiment are also responsive to user actuation in order to cause the playing of the audio file to be paused.
- The correlation of the respective position within the transcripted text report with a corresponding position within the audio file may include utilizing voice recognition to analyze the audio files so as to identify words and comparing the words that are identified from the audio file with words of the transcripted text report to identify a correspondence there between. In this regard, the comparison of the words that are identified from the audio file with words of the transcripted report may include the comparison of a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify a correspondence therebetween.
- In another embodiment, a transcription apparatus is provided that includes processing circuitry configured to receive an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the processing circuitry is configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file. Although the plurality of positions within the transcripted text report may be determined in various manners, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or the commencement of a paragraph. The processing circuitry of this embodiment is also configured to augment the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with a respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- The plurality of selectable elements of one embodiment includes a plurality of icons. In this embodiment, the processing circuitry is configured to augment the transcripted text report by inserting an icon within the transcripted text report at each of the plurality of positions of the transcripted text report. The icons of this embodiment are responsive to user actuation in order to cause the audio file to move to the corresponding position. The icons may also be responsive to user actuation in order to cause the audio file to be played beginning at the corresponding position. In an instance in which the audio file is already being played, the icons may alternatively be responsive to user actuation in order to cause playing of the audio file to be paused.
- The processing circuitry of one embodiment is configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file by utilizing voice recognition to analyze the audio file so as to identify words and then to comparing the words that are identified from the audio file with words of the transcripted text report to identify a correspondence therebetween. The processing circuitry may be configured to compare the words that are identified from the audio file with words from the transcripted text report by comparing a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify a correspondence therebetween.
- In a further embodiment, a computer program product is provided that includes at least one non-transitory computer-readable storage medium having computer-executable program code instructions stored therein. The computer-executable program code instructions include program instructions configured to receive an audio file and a transcripted text report corresponding to the audio file. For each of a plurality of positions within the transcripted text report, the program code instructions of this embodiment are configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file. While a plurality of positions may be determined in various manners, at least some of the plurality of positions within the transcripted text report may be associated with a punctuation mark or commencement of a paragraph. The computer-executable program code instructions of this embodiment also include program code instructions configured to augment the transcripted text report to include a plurality of selectable elements. Each selectable element is associated with the respective position within the transcripted text report. The selectable elements are responsive to user actuation in order to cause the audio file to move to the corresponding position.
- The plurality of selectable elements of one embodiment includes a plurality of icons. In this embodiment, the program code instructions configured to augment the transcripted text report include program code instructions configured to insert an icon within the transcripted text report at each of a plurality of positions of the transcripted text report. The icons may be responsive to user actuation in order to cause the audio file to move to the corresponding position. The icons may also be responsive to user actuation in order to cause the audio file to be played beginning at the corresponding position or to cause playing of the audio file to be paused.
- The program code instructions configured to correlate the respective position within the transcripted text report with a corresponding position within the audio file may, in one embodiment, include program code instructions configured to utilize voice recognition to analyze the audio file to identify words and program code instructions configured to compare the words that are identified from the audio file with words of the transcripted text report to identify correspondence therebetween. The program code instructions that are configured to compare the words that are identified from the audio file with words of a transcripted text report include, in one embodiment, program code instructions configured to compare a collection of words that are identified from the audio file with a sentence from the transcripted text report to identify correspondence therebetween.
- Having thus described certain embodiments of the invention in general terms, reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
-
FIG. 1 is a flow chart illustrating the operations performed by a transcription apparatus configured in accordance with an example embodiment of the present invention; -
FIG. 2 is a block diagram of an transcription apparatus that may be specifically configured in accordance with an example embodiment of the present invention; -
FIG. 3 is an example of a transcripted text report following augmentation in accordance with an example embodiment of the present invention; -
FIG. 4 is an example of the transcripted text report ofFIG. 3 following augmentation in accordance with an example embodiment of the present invention and following selection of an icon within the transcripted text report; and -
FIG. 5 is an example of the transcripted text report ofFIG. 3 following augmentation in accordance with an example embodiment of the present invention and following selection of another icon within the transcripted text report. - The present invention now will be described more fully hereinafter with reference to the accompanying drawings, in which some, but not all embodiments of the inventions are shown. Indeed, these inventions may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will satisfy applicable legal requirements. Like numbers refer to like elements throughout.
- A transcription method, apparatus and compute program product are provided in accordance with example embodiments of the present invention in order to correlate an audio file and a transcripted text report of the audio file. Thus, the transcription method, apparatus and computer program product of an example embodiment may facilitate the review and verification of the transcripted text report by reference to the corresponding audio file. In this regard, the transcripted text report may be reviewed and verified in a more efficient and a quicker manner since those portions of the audio file that correspond to portions of interest within the transcripted text report may be readily identified and reviewed. While the transcription method, apparatus and computer program product may be utilized in conjunction with any type of audio file and any type of transcripted text report, the transcription method, apparatus and computer program product of an example embodiment will be described hereinafter for purposes of example, but not of limitation, in conjunction with the review of an audio file created by a radiologist, such as during or following the review of an image by a radiologist, in order to generate a report that includes the radiologist's findings with respect to the image and that has been verified relative to the audio file.
- In this regard, an audio file may be initially created, such as by dictating to a digital voice recorder or other type of dictation device. The audio file may relate to various subjects including, for example, the reports generated by health care professionals. For example, a radiologist may review an image of a patient and may dictate a report relating to the review of the image, including their findings with respect to the image. The image may be acquired by a number of different imaging modalities including, for example, ultrasound (US), magnetic resonance (MR), positron emission tomography (PET), computed tomography (CT), mammography (MG), digital radiology (DR), computed radiology (CR) or the like.
- The audio file may then be transcribed, such as by a transcriptionist, to create a transcripted text file. In some instances, the transcripted text file is reviewed and verified relative to the corresponding audio file. In this regard, a transcripted text report corresponding to an audio file that represents the report of a health care practitioner, such as report of a radiologist following the review of an image, may be performed to verify that the transcripted text file is an accurate reproduction of the audio file. In this regard, a radiologist, such as a radiologist who initially dictated the audio file or another radiologist, may review the transcripted text file to verify its accuracy, including listening to all or at least portions of the audio file in order to make such a determination.
- In order to improve the efficiency with which the transcripted text may be reviewed in relation to the corresponding audio file, the method of one embodiment may receive an audio file and a transcripted text report corresponding to the audio file, that is, the transcripted text report that represents a transcription of the audio file. See
block 10 ofFIG. 1 . In this regard, the operations represented by the flow chart ofFIG. 1 may be performed by atranscription apparatus 20 that is specifically configured in accordance with embodiments of the present invention. A transcription apparatus may be embodied in any of a variety of computing devices, such as laptop computers, tablet computers, smartphones, desktop computers, computer workstations or the like. In some example embodiments, the transcription apparatus includes various means for performing the various functions described herein. These means may include, for example, one or more of processingcircuitry 22, amemory 24, and/or auser interface 26 for performing the various functions herein described. The means of the apparatus as described herein may be embodied as, for example, circuitry, hardware elements (e.g., a suitably programmed processor, combinational logic circuit, and/or the like), a computer program product comprising computer-readable program instructions (e.g., software or firmware) stored on a computer-readable medium (e.g. memory) that is executable by a suitably configured processing device (e.g., the processing device), or some combination thereof. - The
processing circuitry 22 may, for example, be embodied as various means including one or more microprocessors, one or more coprocessors, one or more multi-core processors, one or more controllers, one or more computers, various other processing elements including integrated circuits such as, for example, an ASIC (application specific integrated circuit) or FPGA (field programmable gate array), or some combination thereof. The processing circuitry may be embodied on a single computing device or may be distributed across a plurality of computing devices collectively configured to function as the apparatus. In some example embodiments, the processing circuitry is configured to execute instructions stored in thememory 24 or otherwise accessible to the processing circuitry. These instructions, when executed by the processing circuitry, may cause thetranscription apparatus 20 to perform one or more of the functionalities of the apparatus as described herein. As such, whether configured by hardware or software methods, or by a combination thereof, the processing circuitry may comprise an entity capable of performing operations according to embodiments of the present invention while configured accordingly. Thus, for example, when the processing circuitry is embodied as an ASIC, FPGA or the like, the processing circuitry may comprise specifically configured hardware for conducting one or more operations described herein. Alternatively, as another example, when the processing circuitry is embodied as an executor of instructions, such as may be stored in the memory, the instructions may specifically configure the processing circuitry to perform one or more algorithms and operations described herein. - The
memory 24 may include, for example, volatile and/or non-volatile memory. Although illustrated inFIG. 2 as a single memory, the memory may comprise a plurality of memories. The plurality of memories may be embodied on a single computing device or distributed across a plurality of computing devices. The memory may comprise, for example, a hard disk, random access memory, cache memory, flash memory, an optical disc (e.g., a compact disc read only memory (CD-ROM), digital versatile disc read only memory (DVD-ROM), or the like), circuitry configured to store information, or some combination thereof. In this regard, the memory may comprise any non-transitory computer readable storage medium. The memory may be configured to store information, data, applications, instructions, or the like for enabling thetranscription apparatus 20 to carry out various functions in accordance with example embodiments of the present invention. For example, in some example embodiments, the memory is configured to buffer input data for processing by theprocessing circuitry 22. Additionally or alternatively, in some example embodiments, the memory is configured to store program instructions for execution by the processing circuitry. - The
user interface 26 may be in communication with theprocessing circuitry 22 to receive an indication of a user input and/or to provide an audible, visual, mechanical, or other output to a user. In one embodiment, the user interface may include, for example, a keyboard and/or a touch sensitive input device, such as a touch sensitive mouse, for receiving user input. The user interface may also include a display for presenting medical images for review and manipulation by a user. The user interface may be in communication with thememory 24, such as via a bus. Although not shown inFIG. 2 , thetranscription apparatus 20 of one embodiment may also include a communication interface, such as one or more antennas, one or more transceivers, etc., for facilitating communication between the transcription apparatus and one or more other computing or communications devices. - Returning now to
FIG. 1 and with reference to thetranscription apparatus 20 of the embodiment ofFIG. 2 , the transcription apparatus may include means, such as theprocessing circuitry 22, the communication interface or the like, for receiving the audio file and the transcripted text report that corresponds to the audio file. In this regard, the audio file and the transcripted text report may be stored inmemory 24 and accessed by the processing circuitry. Alternatively, the transcription apparatus may include a communication interface for receiving the audio file and/or the transcription text report, such as via a wireless connection or a wireline connection. - As shown in
block 12 ofFIG. 1 , thetranscription apparatus 20 may include means, such as theprocessing circuitry 22 or the like, for correlating, for each of a plurality of positions within the transcripted text report, a respective position within the transcripted text report with a corresponding position within the audio file. For example, each word of the transcripted text report may be correlated with a corresponding position within the audio file, each sentence within the transcripted text report may be associated with a corresponding position within the audio file, and/or each paragraph within the transcripted text report may be correlated with a corresponding position within the audio file. Indeed, while the correlation between a respective position within the transcripted text report and a corresponding position within the audio file may be performed on a word basis, a sentence basis or a paragraph basis, as noted above, the various positions within the transcripted text report that are correlated with corresponding positions within the audio file may be identified in other manners, such as by correlating each punctuation mark within the transcripted text report with a corresponding position within the audio file. - The correlation between a respective position within the transcripted text report and a corresponding position within the audio file may be performed by the
processing circuitry 22 in various manners. In one embodiment that is described below for purposes of illustration, but not of limitation, the processing circuitry utilizes voice recognition to analyze the audio file so as to identify words. In this regard, the processing circuitry may implement a voice recognition engine that is configured to recognize and to identify the pronunciation of a plurality of words and, in one embodiment, that may be trained so as to recognize the pronunciation of a plurality of words, such as a plurality of the most frequently used words, by a particular user, such as a radiologist. - In one embodiment, the
processing circuitry 22 and, more particularly, the voice recognition engine may identify each unique word that is included within the audio file. For example, if the audio file includes a following sentence, “I am working on it, and I am close.” The processing circuitry, such as a voice recognition engine of the processing circuitry, may identify “I”, “am”, “working”, “on”, “it” and “close” as the unique words that are included within the audio file. The processing circuitry and, more particularly, the voice recognition engine implemented by the processing circuitry, also identifies the position of each word within the audio file. In this regard, the position of a word within the audio file will be hereinafter referenced as the audio position of the word. While the position within the audio file may be identified in various manners, the processing circuitry and, more particularly, the voice recognition engine of one embodiment may identify the audio position of each word based upon the elapsed time from the beginning of the audio file to the word as the audio position of the word in one embodiment. - In the foregoing example in which the word begins “I” appears on two occasions, the
processing circuitry 22 and, more particularly, the voice recognition engine implemented by the processing circuitry, may identify both audio positions of the word “I”, such as at one millisecond and five milliseconds relative to the beginning of the audio file. As such, in an instance in which a word appears a plurality of times within an audio file, the processing circuitry, such as a voice recognition engine implemented by the processing circuitry, may identify the position of each occurrence of the word within the audio file. Although the position of a respective word that appears multiple times may be represented as a compact form, such as [“I” 1 ms, 5 ms], the compact representation may be expanded such that each occurrence of the same word is individually associated with a corresponding position within the audio file, such as [“I” 1 ms] and [“I” 5 ms]. In this regard, the association of a word and its corresponding audio location may be referred to as a word and audio location pair, with a plurality of word and audio location pairs, such as all of the word and audio location pairs that appear within the audio file, being referred to as a collection. Based upon the respective audio locations, the collection of word and audio location pairs may be placed in an order of occurrence. - In one embodiment, the
processing circuitry 22, such as the voice recognition engine implemented by the processing circuitry, may compare subsequences from the collection of word and audio location pairs that have been identified from the audio file with sentences from the transcripted text report on a sentence-by-sentence basis in order to determine a correspondence therebetween. For each sentence of the transcripted text report, the processing circuitry, such as the voice recognition engine implemented by the processing circuitry of one embodiment, may identify the longest subsequence within the collection of words and audio location pairs that matches the respective sentence of the transcripted text report and may identify the resulting subsequence of word and audio location pairs to correspond to the respective sentence. - The
transcription apparatus 20 and, more particularly, theprocessing circuitry 22, thememory 24 or the like may also augment the transcripted text report to include a plurality of selectable elements. Seeblock 14 ofFIG. 1 . In this regard, each selectable element is associated with a respective position within the transcripted text report. Although selectable elements may be associated with positions within a transcripted text report that are defined in various manners, the processing circuitry of one embodiment may be configured to associate a selectable element with the beginning of each paragraph and/or with each punctuation mark within the transcripted text report. In addition to being associated with a respective position within the transcripted text report, the selectable element or at least the respective position within the transcripted text report with which the selectable element is associated may also be associated with a corresponding position within the audio file based upon the correlation of the respective position within the transcripted text report and the corresponding position within the audio file. Thus, each selectable element that is associated with the beginning of a paragraph from the transcripted text report may also be associated with the corresponding position within the audio file that is correlated to the beginning of the respective paragraph. Further, each selectable element that is associated with a punctuation mark within the transcripted text report may also be associated with the corresponding position within the audio file that is correlated to the respective punctuation mark. As such, a user who is reviewing the transcripted text report, following its augmentation, may actuate a selectable element in order to cause the audio file to move to the corresponding position, that is, to move to the position within the audio file that has been correlated with the position within the transcripted text report with which the selectable element is associated. - Various types of selectable elements may be utilized in order to augment a transcripted text report. For example, the first word of a paragraph may serve as the selectable element associated with the beginning of the paragraph, while the punctuation mark may serve as the selectable element associated with itself. In one embodiment, however, a transcripted text report may be augmented with a plurality of icons that serve as the selectable elements. In this embodiment, the transcripted text report may be augmented by inserting an icon within the transcripted text report at each of the plurality of positions of the transcripted text report that are to be associated with selectable elements. By way of example,
FIG. 3 illustrates an initial portion of a speech by President Obama that is entitled “Yes We Can”. As shown, an icon has been associated with the beginning of each paragraph as well as with each punctuation mark. By selecting any one of the icons, such as by double clicking upon a respective icon, the transcription method and apparatus of one embodiment will cause the audio file to be moved to the corresponding position that has been correlated with the position within the transcripted text report that is associated with the icon that has been selected. Thus, selection of the icon designated 30 a inFIG. 3 , will cause the audio file to move to a position designated 1:22 as shown by thetime scroll bar 32 ofFIG. 4 . Alternatively, selection of the icon designated 30 b inFIG. 3 will cause the audio file to move to the position designated 1:54 as shown by thetime scroll bar 32 of inFIG. 5 . Thus, the inclusion of icons or other selectable elements within the transcripted text report permits a user to quickly and accurately move to different portions of the audio file in order to listen to the portion of the audio file that corresponds to the portion of the transcripted text report associated with the icon that has been selected. - In an instance in which the audio file is not currently being played, the selection of an icon, such as an icon as shown in
FIG. 3 , not only causes the audio file to move to the position associated with the selected icon, but also to commence playing of the audio file at the selected position. In order to cause the audio file to stop playing, the user may then click any of the icons as shown inFIGS. 4 and 5 . As such, the icons may include a representation of the function, e.g., play, pause, etc., associated with the icon. In this regard, in an instance in which the audio file is not being played, theicons 30 may include a representation of a “play” button as shown inFIG. 3 so as to provide a visual indication to the user regarding the associated function that is initiated upon selection of the icon. However, once the audio file begins to play, the representation of theicons 34 may be changed as shown inFIGS. 4 and 5 to include a representation of a “pause” button so as to provide a visual indication to the user regarding the function that would be initiated upon actuation of the icon. As such, the selection of an icon in an instance in which the audio file is being played and the selection of an icon at the same position within the transcripted text report in an instance in which the audio file is not being played may initiate different functions, such as with the function brought about by actuation of the icon alternately toggling between playing of the selected portion of the audio file in an instance in which the audio file is not currently being played and ceasing further playing of the audio file in an instance in which the audio file is currently being played. In one embodiment, in an instance in which the audio file is being played, theicons 34 may also serve as a mechanism to permit the volume to be readily adjusted. In this regard, a user may place the cursor in alignment with a respective icon and may then provide an input, such as rotation of a scroll wheel carried by a computer mouse, in order to controllably adjust the volume of the audio file being played. - As such, a user may efficiently navigate through the transcripted text file and may review the corresponding portions of the audio file without having to guess as to the relevant portions of the audio file. Accordingly, in an embodiment in which a radiologist is utilizing the transcription method, apparatus and computer program product of an example embodiment of the present invention to review the accuracy of a transcripted text file, the radiologist may review the transcripted text file in conjunction with the audio file and may quickly and efficiently replay any selected portions of the audio file, such as those portions of the audio file that capture the key findings of the radiologist. As such, the transcription method, apparatus and computer program product may improve the overall efficiency of the report dictation, transcription and verification process.
- As described above,
FIG. 1 illustrates a flowchart of a system, method, and computer program product according to example embodiments of the invention. It will be understood that each block of the flowchart, and combinations of blocks in the flowchart, may be implemented by various means, such as hardware and/or a computer program product comprising one or more computer-readable mediums having computer readable program instructions stored thereon. For example, one or more of the procedures described herein may be embodied by computer program instructions of a computer program product. In this regard, the computer program product(s) which embody the procedures described herein may be stored by one ormore memory devices 24 of a computing device and executed by processingcircuitry 22. In some embodiments, the computer program instructions comprising the computer program product(s) which embody the procedures described above may be stored by memory devices of a plurality of computing devices. As will be appreciated, any such computer program product may be loaded onto a computer or other programmable apparatus to produce a machine, such that the computer program product including the instructions which execute on the computer or other programmable apparatus creates means for implementing the functions specified in the flowchart block(s). Further, the computer program product may comprise one or more computer-readable memories on which the computer program instructions may be stored such that the one or more computer-readable memories can direct a computer or other programmable apparatus to function in a particular manner, such that the computer program product comprises an article of manufacture which implements the function specified in the flowchart block(s). The computer program instructions of one or more computer program products may also be loaded onto a computer or other programmable apparatus to cause a series of operations to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions which execute on the computer or other programmable apparatus implement the functions specified in the flowchart block(s). - Accordingly, blocks or steps of the flowchart support combinations of means for performing the specified functions and combinations of steps for performing the specified functions. It will also be understood that one or more blocks of the flowchart, and combinations of blocks in the flowchart, may be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or combinations of special purpose hardware and computer program product(s).
- The above described functions may be carried out in many ways. For example, any suitable means for carrying out each of the functions described above may be employed to carry out embodiments of the invention. In one embodiment, suitably configured
processing circuitry 22 may provide all or a portion of the elements of the invention. In another embodiment, all or a portion of the elements of the invention may be configured by and operate under control of a computer program product. The computer program product for performing the methods of embodiments of the invention includes a computer-readable storage medium, such as the non-volatile storage medium, and computer-readable program code portions, such as a series of computer instructions, embodied in the computer-readable storage medium. - Many modifications and other embodiments of the inventions set forth herein will come to mind to one skilled in the art to which these inventions pertain having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the inventions are not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/538,671 US9805118B2 (en) | 2012-06-29 | 2012-06-29 | Transcription method, apparatus and computer program product |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/538,671 US9805118B2 (en) | 2012-06-29 | 2012-06-29 | Transcription method, apparatus and computer program product |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140006020A1 true US20140006020A1 (en) | 2014-01-02 |
US9805118B2 US9805118B2 (en) | 2017-10-31 |
Family
ID=49779003
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/538,671 Active 2034-04-19 US9805118B2 (en) | 2012-06-29 | 2012-06-29 | Transcription method, apparatus and computer program product |
Country Status (1)
Country | Link |
---|---|
US (1) | US9805118B2 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170161382A1 (en) * | 2015-12-08 | 2017-06-08 | Snapchat, Inc. | System to correlate video data and contextual data |
US20220270610A1 (en) * | 2019-07-15 | 2022-08-25 | Axon Enterprise, Inc. | Methods and systems for transcription of audio data |
US11568231B2 (en) * | 2017-12-08 | 2023-01-31 | Raytheon Bbn Technologies Corp. | Waypoint detection for a contact center analysis system |
US11822600B2 (en) | 2015-09-15 | 2023-11-21 | Snap Inc. | Content tagging |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180270350A1 (en) | 2014-02-28 | 2018-09-20 | Ultratec, Inc. | Semiautomated relay method and apparatus |
US10389876B2 (en) | 2014-02-28 | 2019-08-20 | Ultratec, Inc. | Semiautomated relay method and apparatus |
US10748523B2 (en) | 2014-02-28 | 2020-08-18 | Ultratec, Inc. | Semiautomated relay method and apparatus |
US20180034961A1 (en) | 2014-02-28 | 2018-02-01 | Ultratec, Inc. | Semiautomated Relay Method and Apparatus |
US10878721B2 (en) | 2014-02-28 | 2020-12-29 | Ultratec, Inc. | Semiautomated relay method and apparatus |
US11539900B2 (en) | 2020-02-21 | 2022-12-27 | Ultratec, Inc. | Caption modification and augmentation systems and methods for use by hearing assisted user |
Citations (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6345253B1 (en) * | 1999-04-09 | 2002-02-05 | International Business Machines Corporation | Method and apparatus for retrieving audio information using primary and supplemental indexes |
US20070011012A1 (en) * | 2005-07-11 | 2007-01-11 | Steve Yurick | Method, system, and apparatus for facilitating captioning of multi-media content |
US7231351B1 (en) * | 2002-05-10 | 2007-06-12 | Nexidia, Inc. | Transcript alignment |
US20080122796A1 (en) * | 2006-09-06 | 2008-05-29 | Jobs Steven P | Touch Screen Device, Method, and Graphical User Interface for Determining Commands by Applying Heuristics |
US20080165153A1 (en) * | 2007-01-07 | 2008-07-10 | Andrew Emilio Platzer | Portable Multifunction Device, Method, and Graphical User Interface Supporting User Navigations of Graphical Objects on a Touch Screen Display |
US20080168349A1 (en) * | 2007-01-07 | 2008-07-10 | Lamiraux Henri C | Portable Electronic Device, Method, and Graphical User Interface for Displaying Electronic Documents and Lists |
US20080220752A1 (en) * | 2007-01-07 | 2008-09-11 | Scott Forstall | Portable Multifunction Device, Method, and Graphical User Interface for Managing Communications Received While in a Locked State |
US7444285B2 (en) * | 2002-12-06 | 2008-10-28 | 3M Innovative Properties Company | Method and system for sequential insertion of speech recognition results to facilitate deferred transcription services |
US20080320391A1 (en) * | 2007-06-20 | 2008-12-25 | Lemay Stephen O | Portable Multifunction Device, Method, and Graphical User Interface for Playing Online Videos |
US20090007017A1 (en) * | 2007-06-29 | 2009-01-01 | Freddy Allen Anzures | Portable multifunction device with animated user interface transitions |
US20090021482A1 (en) * | 2007-07-20 | 2009-01-22 | Ying-Chu Lee | Virtually multiple wheels and method of manipulating multifunction tool icons thereof |
US20090112467A1 (en) * | 2007-10-31 | 2009-04-30 | Ning Jiang | Map-centric service for social events |
US20090307615A1 (en) * | 2008-06-06 | 2009-12-10 | Julien Jalon | Browsing or searching user interfaces and other aspects |
US20100100470A1 (en) * | 2008-10-16 | 2010-04-22 | Bank Of America Corporation | Financial planning tool |
US7836412B1 (en) * | 2004-12-03 | 2010-11-16 | Escription, Inc. | Transcription editing |
US20100325043A1 (en) * | 2008-10-16 | 2010-12-23 | Bank Of America Corporation | Customized card-building tool |
US20110087985A1 (en) * | 2008-10-16 | 2011-04-14 | Bank Of America Corporation | Graph viewer |
US20110107265A1 (en) * | 2008-10-16 | 2011-05-05 | Bank Of America Corporation | Customizable graphical user interface |
US20110125847A1 (en) * | 2009-11-25 | 2011-05-26 | Altus Learning System, Inc. | Collaboration networks based on user interactions with media archives |
US20110167350A1 (en) * | 2010-01-06 | 2011-07-07 | Apple Inc. | Assist Features For Content Display Device |
US20110239107A1 (en) * | 2010-03-29 | 2011-09-29 | Phillips Michael E | Transcript editor |
US20120200774A1 (en) * | 2011-02-07 | 2012-08-09 | Ehlers Sr Gregory Allen | Audio and video distribution system with latency delay compensator |
US20120290974A1 (en) * | 2011-01-20 | 2012-11-15 | Vibrant Media, Inc. | Systems and methods for providing a discover prompt to augmented content of a web page |
US20120297429A1 (en) * | 2011-05-17 | 2012-11-22 | Cbs Interactive, Inc. | Emulating Television Viewing Experience In A Browser |
US20130058471A1 (en) * | 2011-09-01 | 2013-03-07 | Research In Motion Limited. | Conferenced voice to text transcription |
US20130290902A1 (en) * | 2010-12-30 | 2013-10-31 | Electrolux Home Products, Inc. | User control interface for an appliance, and associated method |
US20140327677A1 (en) * | 2012-01-06 | 2014-11-06 | Thomson Licensing | Method and system for providing a graphical representation on a second screen of social messages related to content on a first screen |
US9639623B2 (en) * | 2012-12-20 | 2017-05-02 | Cable Television Laboratories, Inc. | Method for remotely controlling a video playing within a first web page at a first device from a second device using a document object model |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE60202453T2 (en) | 2001-03-29 | 2006-01-19 | Koninklijke Philips Electronics N.V. | SYNCHRONIZING AN AUDIO AND TEXT CURSOR DURING EDITING |
-
2012
- 2012-06-29 US US13/538,671 patent/US9805118B2/en active Active
Patent Citations (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6345253B1 (en) * | 1999-04-09 | 2002-02-05 | International Business Machines Corporation | Method and apparatus for retrieving audio information using primary and supplemental indexes |
US7231351B1 (en) * | 2002-05-10 | 2007-06-12 | Nexidia, Inc. | Transcript alignment |
US7444285B2 (en) * | 2002-12-06 | 2008-10-28 | 3M Innovative Properties Company | Method and system for sequential insertion of speech recognition results to facilitate deferred transcription services |
US7836412B1 (en) * | 2004-12-03 | 2010-11-16 | Escription, Inc. | Transcription editing |
US20070011012A1 (en) * | 2005-07-11 | 2007-01-11 | Steve Yurick | Method, system, and apparatus for facilitating captioning of multi-media content |
US20080122796A1 (en) * | 2006-09-06 | 2008-05-29 | Jobs Steven P | Touch Screen Device, Method, and Graphical User Interface for Determining Commands by Applying Heuristics |
US20080168349A1 (en) * | 2007-01-07 | 2008-07-10 | Lamiraux Henri C | Portable Electronic Device, Method, and Graphical User Interface for Displaying Electronic Documents and Lists |
US20080220752A1 (en) * | 2007-01-07 | 2008-09-11 | Scott Forstall | Portable Multifunction Device, Method, and Graphical User Interface for Managing Communications Received While in a Locked State |
US20080165153A1 (en) * | 2007-01-07 | 2008-07-10 | Andrew Emilio Platzer | Portable Multifunction Device, Method, and Graphical User Interface Supporting User Navigations of Graphical Objects on a Touch Screen Display |
US20080320391A1 (en) * | 2007-06-20 | 2008-12-25 | Lemay Stephen O | Portable Multifunction Device, Method, and Graphical User Interface for Playing Online Videos |
US20090007017A1 (en) * | 2007-06-29 | 2009-01-01 | Freddy Allen Anzures | Portable multifunction device with animated user interface transitions |
US20090021482A1 (en) * | 2007-07-20 | 2009-01-22 | Ying-Chu Lee | Virtually multiple wheels and method of manipulating multifunction tool icons thereof |
US20090112467A1 (en) * | 2007-10-31 | 2009-04-30 | Ning Jiang | Map-centric service for social events |
US20090307615A1 (en) * | 2008-06-06 | 2009-12-10 | Julien Jalon | Browsing or searching user interfaces and other aspects |
US20110087985A1 (en) * | 2008-10-16 | 2011-04-14 | Bank Of America Corporation | Graph viewer |
US20100100470A1 (en) * | 2008-10-16 | 2010-04-22 | Bank Of America Corporation | Financial planning tool |
US20110107265A1 (en) * | 2008-10-16 | 2011-05-05 | Bank Of America Corporation | Customizable graphical user interface |
US20100325043A1 (en) * | 2008-10-16 | 2010-12-23 | Bank Of America Corporation | Customized card-building tool |
US20110125847A1 (en) * | 2009-11-25 | 2011-05-26 | Altus Learning System, Inc. | Collaboration networks based on user interactions with media archives |
US20110167350A1 (en) * | 2010-01-06 | 2011-07-07 | Apple Inc. | Assist Features For Content Display Device |
US20110239107A1 (en) * | 2010-03-29 | 2011-09-29 | Phillips Michael E | Transcript editor |
US20130290902A1 (en) * | 2010-12-30 | 2013-10-31 | Electrolux Home Products, Inc. | User control interface for an appliance, and associated method |
US20120290974A1 (en) * | 2011-01-20 | 2012-11-15 | Vibrant Media, Inc. | Systems and methods for providing a discover prompt to augmented content of a web page |
US20120200774A1 (en) * | 2011-02-07 | 2012-08-09 | Ehlers Sr Gregory Allen | Audio and video distribution system with latency delay compensator |
US20120297429A1 (en) * | 2011-05-17 | 2012-11-22 | Cbs Interactive, Inc. | Emulating Television Viewing Experience In A Browser |
US20130058471A1 (en) * | 2011-09-01 | 2013-03-07 | Research In Motion Limited. | Conferenced voice to text transcription |
US20140327677A1 (en) * | 2012-01-06 | 2014-11-06 | Thomson Licensing | Method and system for providing a graphical representation on a second screen of social messages related to content on a first screen |
US9639623B2 (en) * | 2012-12-20 | 2017-05-02 | Cable Television Laboratories, Inc. | Method for remotely controlling a video playing within a first web page at a first device from a second device using a document object model |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11822600B2 (en) | 2015-09-15 | 2023-11-21 | Snap Inc. | Content tagging |
US20170161382A1 (en) * | 2015-12-08 | 2017-06-08 | Snapchat, Inc. | System to correlate video data and contextual data |
US12216702B1 (en) | 2015-12-08 | 2025-02-04 | Snap Inc. | Redirection to digital content based on image-search |
US11568231B2 (en) * | 2017-12-08 | 2023-01-31 | Raytheon Bbn Technologies Corp. | Waypoint detection for a contact center analysis system |
US20220270610A1 (en) * | 2019-07-15 | 2022-08-25 | Axon Enterprise, Inc. | Methods and systems for transcription of audio data |
US11640824B2 (en) * | 2019-07-15 | 2023-05-02 | Axon Enterprise, Inc. | Methods and systems for transcription of audio data |
US12062374B2 (en) | 2019-07-15 | 2024-08-13 | Axon Enterprise, Inc. | Methods and systems for transcription of audio data |
Also Published As
Publication number | Publication date |
---|---|
US9805118B2 (en) | 2017-10-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9805118B2 (en) | Transcription method, apparatus and computer program product | |
US11650732B2 (en) | Method and system for generating transcripts of patient-healthcare provider conversations | |
US11894140B2 (en) | Interface for patient-provider conversation and auto-generation of note or summary | |
JP6078057B2 (en) | Document expansion in dictation-based document generation workflow | |
Aldaz et al. | Hands-free image capture, data tagging and transfer using Google Glass: a pilot study for improved wound care management | |
US20230335261A1 (en) | Combining natural language understanding and image segmentation to intelligently populate text reports | |
US20190272921A1 (en) | Automated Diagnostic Support System for Clinical Documentation Workflows | |
US20100010806A1 (en) | Storage system for symptom information of Traditional Chinese Medicine (TCM) and method for storing TCM symptom information | |
US11830607B2 (en) | Systems and methods for facilitating image finding analysis | |
CN112714914B (en) | System and method for memory augmentation domain adaptation | |
KR102436787B1 (en) | Method, electronic device and program for estimating annotator and ananlyzing anotation using a plurarity of traning models | |
US20150066528A1 (en) | Computer-implemented method and system for generating a report | |
US20230118299A1 (en) | Radiologist fingerprinting | |
US20100023312A1 (en) | System and method enabling bi-translation for improved prescription accuracy | |
EP4010877A1 (en) | Method and systems for the automated detection of free fluid using artificial intelligence for the focused assessment sonography for trauma ("fast") examination for trauma care | |
KR102292041B1 (en) | Method for comprising time interval unit data set to predict disease and device thereof | |
EP4498379A1 (en) | Medical image analysis | |
JPWO2020084101A5 (en) | ||
WO2022192893A1 (en) | Artificial intelligence system and method for generating medical impressions from text-based medical reports | |
WO2025022176A1 (en) | Medical image analysis | |
Enrico et al. | The digital scribe | |
Suominen et al. | Clinical speech to text: Evaluation setting | |
JP2007172539A (en) | Word retrieval device and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MCKESSON FINANCIAL HOLDINGS, BERMUDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KO, TAK M.;ZIGIC, DRAGAN;SIGNING DATES FROM 20120628 TO 20120629;REEL/FRAME:028472/0804 |
|
AS | Assignment |
Owner name: MCKESSON FINANCIAL HOLDINGS UNLIMITED COMPANY, BERMUDA Free format text: CHANGE OF NAME;ASSIGNOR:MCKESSON FINANCIAL HOLDINGS;REEL/FRAME:041329/0879 Effective date: 20161130 Owner name: MCKESSON FINANCIAL HOLDINGS UNLIMITED COMPANY, BER Free format text: CHANGE OF NAME;ASSIGNOR:MCKESSON FINANCIAL HOLDINGS;REEL/FRAME:041329/0879 Effective date: 20161130 |
|
AS | Assignment |
Owner name: MCKESSON CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MCKESSON FINANCIAL HOLDINGS UNLIMITED COMPANY;REEL/FRAME:041355/0408 Effective date: 20161219 |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH CAROLINA Free format text: SECURITY AGREEMENT;ASSIGNORS:CHANGE HEALTHCARE HOLDINGS, LLC;CHANGE HEALTHCARE, INC.;CHANGE HEALTHCARE HOLDINGS, INC.;AND OTHERS;REEL/FRAME:041858/0482 Effective date: 20170301 Owner name: BANK OF AMERICA, N.A., AS COLLATERAL AGENT, NORTH Free format text: SECURITY AGREEMENT;ASSIGNORS:CHANGE HEALTHCARE HOLDINGS, LLC;CHANGE HEALTHCARE, INC.;CHANGE HEALTHCARE HOLDINGS, INC.;AND OTHERS;REEL/FRAME:041858/0482 Effective date: 20170301 |
|
AS | Assignment |
Owner name: PF2 IP LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MCKESSON CORPORATION;REEL/FRAME:041938/0501 Effective date: 20170301 |
|
AS | Assignment |
Owner name: CHANGE HEALTHCARE LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PF2 IP LLC;REEL/FRAME:041966/0356 Effective date: 20170301 |
|
AS | Assignment |
Owner name: CHANGE HEALTHCARE LLC, GEORGIA Free format text: CHANGE OF ADDRESS;ASSIGNOR:CHANGE HEALTHCARE LLC;REEL/FRAME:042082/0061 Effective date: 20170323 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: CHANGE HEALTHCARE HOLDINGS, LLC, TENNESSEE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHANGE HEALTHCARE LLC;REEL/FRAME:046449/0899 Effective date: 20180414 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: CHANGE HEALTHCARE HOLDINGS, LLC, MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE TECHNOLOGIES, LLC (FORMERLY KNOWN AS MCKESSON TECHNOLOGIES LLC), MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE HOLDINGS, INC., MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE OPERATIONS, LLC, MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE PERFORMANCE, INC. (FORMERLY KNOWN AS CHANGE HEALTHCARE, INC.), MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE SOLUTIONS, LLC, MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 Owner name: CHANGE HEALTHCARE RESOURCES, LLC (FORMERLY KNOWN AS ALTEGRA HEALTH OPERATING COMPANY LLC), MINNESOTA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:061620/0054 Effective date: 20221003 |