US20080082334A1 - Multi-pass speech analytics - Google Patents
Multi-pass speech analytics Download PDFInfo
- Publication number
- US20080082334A1 US20080082334A1 US11/924,201 US92420107A US2008082334A1 US 20080082334 A1 US20080082334 A1 US 20080082334A1 US 92420107 A US92420107 A US 92420107A US 2008082334 A1 US2008082334 A1 US 2008082334A1
- Authority
- US
- United States
- Prior art keywords
- communications
- tier
- analytics
- event
- speech recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000004891 communication Methods 0.000 claims abstract description 156
- 238000000034 method Methods 0.000 claims abstract description 65
- 230000008569 process Effects 0.000 claims abstract description 47
- 238000012545 processing Methods 0.000 claims description 14
- 239000003795 chemical substances by application Substances 0.000 description 25
- 238000006243 chemical reaction Methods 0.000 description 19
- 238000004458 analytical method Methods 0.000 description 12
- 238000012517 data analytics Methods 0.000 description 12
- 238000005516 engineering process Methods 0.000 description 11
- 238000010586 diagram Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 10
- 230000003993 interaction Effects 0.000 description 10
- 238000013515 script Methods 0.000 description 9
- 230000008451 emotion Effects 0.000 description 8
- 238000005457 optimization Methods 0.000 description 6
- 238000011156 evaluation Methods 0.000 description 5
- 238000007726 management method Methods 0.000 description 5
- 238000012544 monitoring process Methods 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 238000000605 extraction Methods 0.000 description 3
- 238000001914 filtration Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000013439 planning Methods 0.000 description 3
- 238000012795 verification Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000003909 pattern recognition Methods 0.000 description 2
- 238000000275 quality assurance Methods 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 238000003339 best practice Methods 0.000 description 1
- 239000000872 buffer Substances 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000013523 data management Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000037406 food intake Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 208000018910 keratinopathic ichthyosis Diseases 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 238000001303 quality assessment method Methods 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/32—Multiple recognisers used in sequence or in parallel; Score combination systems therefor, e.g. voting systems
Definitions
- call center infrastructures in which a customer or other user can call to receive information related to the business.
- the customer may be connected with a customer service representative to provide the desired information.
- the customer may be connected with different customer service representatives.
- the customer may be provided with varying levels of customer service with respect to the interaction with the customer service representative. Because most businesses desire to provide the highest possible quality of customer service, many businesses have turned to recording the communication between the customer and the customer service representative. While recording this data has proven beneficial in many cases, many businesses receive call volumes that inhibit the business from reviewing all of the call data received.
- LVCSR large vocabulary continuous speech recognition
- At least one embodiment of a method includes receiving audio data associated with first communications desired to be analyzed for an event, performing a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and performing a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- LVCSR large vocabulary continuous speech recognition
- At least one embodiment of a system includes an interface configured to receive audio data associated with first communications desired to be analyzed for an event, and a processor coupled to the interface and configured to perform a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and perform a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- LVCSR large vocabulary continuous speech recognition
- At least one embodiment includes a computer readable medium having stored thereon instructions that, when executed by a processor, direct the processor to receive audio data associated with first communications desired to be analyzed for an event, perform a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and perform a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- LVCSR large vocabulary continuous speech recognition
- FIG. 1 is an exemplary embodiment of a system in which a user may be provided with customer service.
- FIG. 2 is an exemplary embodiment of a recording network configuration, in which recording and analytics may be provided, similar to the network configuration from FIG. 1 .
- FIG. 3 is a functional block diagram illustrating an exemplary embodiment of an analytic component, similar to the analytic component from FIG. 2 .
- FIG. 4 is an exemplary embodiment of a functional flow diagram, illustrating components that may be utilized for a phonetics based analytic infrastructure, as may be utilized in the network of FIG. 1 .
- FIG. 5 is an exemplary embodiment of a functional block diagram illustrating speech recognition according to a text speech engine, similar to the diagram from FIG. 4 .
- FIG. 6 is a network diagram illustrating an exemplary embodiment of a multi-tier speech recognition configuration, similar to the network configuration from FIG. 1 .
- FIG. 7 is a network configuration illustrating an exemplary embodiment of a multi-tier speech recognition configuration with a plurality of servers, similar to the network configuration from FIG. 6 .
- FIG. 8 is a network configuration illustrating an exemplary embodiment of a three-tier speech recognition configuration, similar to the network configuration from FIG. 7 .
- FIG. 9A is a flowchart illustrating the exemplary embodiment of a process that may be utilized to perform speech recognition functionality in a multi-tier speech recognition environment, such as the environment from FIG. 6 .
- FIG. 9B is a continuation from the flowchart from FIG. 9A .
- FIG. 10A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing speech recognition in a first tier upon a determination that utilization of a second speech recognition tier is desired, similar to the flowchart from FIGS. 9A and 9B .
- FIG. 10B is a continuation of the flowchart from FIG. 10A .
- FIG. 11A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing a user option for providing an agent an option to send audio data to a second tier speech recognition system, similar to the flowchart from FIGS. 10A and 10B .
- FIG. 11B is a continuation of the flowchart from FIG. 11A .
- FIG. 12A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for determining from a plurality of second tier speech recognition servers, which server tier direct audio data, similar to the flowchart from FIGS. 11A and 11B .
- FIG. 12B is a continuation of the flowchart from FIG. 12A .
- first tier speech to text analytics and second tier speech to text analytics are used.
- a first tier may be configured for speech to text conversion and a second tier may be configured for speech to text analytics.
- Other embodiments are also included, as discussed with reference to the drawings.
- FIG. 1 is a nonlimiting example of a system in which a user may be provided with customer service.
- a caller can initiate a communication request from a communications device 104 .
- the communications request can be sent to call center 106 via network 100 .
- Network 100 may include a Public Switched Telephone Network (PSTN), an Internet Protocol (IP) network, a cellular network, an Integrated Services Digital Network (ISDN), a Time Division Multiplexing (TDM) network, and/or other networks.
- PSTN Public Switched Telephone Network
- IP Internet Protocol
- ISDN Integrated Services Digital Network
- TDM Time Division Multiplexing
- a user at communications device 104 may facilitate a communication with a customer service representative. Additionally, during the communication, the communication can be recorded, such that a business associated with call center 106 can determine the quality of customer service provided to the user of communications device 104 .
- an audio recording can be provided to an analyst to determine the quality of customer service
- some embodiments may include a text to voice conversion of the communication.
- Large Vocabulary Continuous Speech Recognition (LVCSR) may be utilized to create an English translated textual document associated with the communication. While an LVCSR speech recognized textual document may provide enhanced searching capabilities related to the communication, depending on an accuracy threshold, LVCSR technologies may be slow in execution. Similarly, in many phonetic technologies for speech recognition, processing of search functions associated with the communication may be slow.
- a user can send a communication request via communication device 104
- some embodiments may provide that a user utilizing computing device 108 may initiate a communication to call center 106 via network 100 .
- a user may utilize a soft phone and/or other communications logic provided for initiating and facilitating a communication.
- a call center can include, but is not limited to, outsourced contact centers, outsourced customer relationship management, customer relationship management, voice of the customer, customer interaction, contact center, multi-media contact center, remote office, distributed enterprise, work-at-home agents, remote agents, branch office, back office, performance optimization, workforce optimization, hosted contact centers, and speech analytics, for example.
- FIG. 2 is a nonlimiting example of a system for providing recording and analytics, similar to the network configuration from FIG. 1 .
- communications device 104 can facilitate communications with call center 106 via network 100 .
- Call center 106 can include one or more elements, such as a recorder subsystem component 204 , an extraction filtering component 206 , and a speech processing firmware component 208 .
- Also included in call center 106 is an advanced data analytics (pattern recognition) component 218 and an atlas component 210 . Included in the atlas component 210 is a speech package 212 , a desktop event 214 , and an interactions package 216 .
- Call center 106 may also include an analytic scorecard 220 , a quality management (QM) evaluations component 222 , and enterprise reporting component 224 , and a speech and replay component 226 .
- An agent 228 can utilize one or more of the components of call center 106 to facilitate a communication with a caller on communications device 104 .
- an analyst 230 can utilize one or more components of call center 106 to analyze the quality of the communications between the agent 228 and the caller associated with communications device 104 .
- a supervisor 232 may also have access to components of call center 106 to oversee the agent 228 and/or the analyst 230 and their interactions with a caller on communications device 104 .
- a recognition engine cluster 202 may be coupled to call center 106 directly and/or via network 100 .
- Recognition engine cluster 202 may include one or more servers that may provide speech recognition functionality to call center 106 .
- a communication between a caller on communications device 104 and an agent 228 , via network 100 may first be received by a recorder subsystem component 204 .
- Recorder subsystem component 204 may record the communications in an audio format.
- the recorder audio may then be sent to an extraction filtering component 206 which may be configured to extract the dialogue (e.g., remove noise and other unwanted sounds) from the recording.
- the recorded communication can then be sent to a speech-processing framework component 208 for converting the recorded audio communication into a textual format.
- Conversion of the audio into a textual format may be facilitated by a recognition engine cluster 202 , however this is not a requirement. Regardless, conversion from the audio format to a textual format may be facilitated via LVCSR speech recognition technologies and/or phonetic speech recognition technologies, as discussed in more detail below.
- Advanced data analytics component 218 may be configured to provide analysis associated with the speech to text converted communication to determine the quality of customer service provided to the caller of communications device 104 .
- Advanced data analytics component 218 may utilize atlas component 210 for facilitation of this analysis. More specifically, atlas component 210 may include a speech package component 212 that may be configured to analyze various patterns in the speech of the caller of communications device 104 .
- desktop event component 214 may be configured to analyze one or more actions that the user of communications device takes on their communications device 104 . More specifically, a network 100 may facilitate communications in an IP network.
- communications device 104 may facilitate both audio and/or data communications that may include audio, video, images, and/or other data.
- advanced data analytics component 218 may utilize an actions package 216 to determine various components of the interaction between agent 228 and the caller of communications device 104 . Advanced data analytics component 218 may then make a determination based on predetermined criteria of the quality of call service provided by agent 220 .
- Advanced data analytics component 218 may then facilitate creation of an analytic scorecard 220 and provide enterprise reporting 224 . Additionally, call center may provide quality management evaluations 222 , as well as speech and replay communications 226 . This data may be viewed by an agent 228 , an analyst 230 , and/or a supervisor 232 . Additionally, as discussed in more detail below, an analyst 230 may further analyze the data to provide a basis for advanced data analytics component 218 to determine the quality of customer service.
- FIG. 3 is a functional block diagram illustrating an analytic component, similar to the analytic component from FIG. 2 .
- advanced data analytics component 218 includes a processor 382 , volatile and nonvolatile memory 384 , a display interface 394 , data storage 395 , and one or more input and/or output (I/O) device interface(s) 396 that are communicatively coupled via a local interface 392 .
- the local interface 392 can include, for example but not limited to, one or more buses and/or other wired or wireless connections.
- the local interface 392 may have additional elements, which are omitted for simplicity, such as controllers, buffers (caches), drivers, repeaters, and receivers to enable communications. Further, the local interface may include address, control, and/or data connections to enable appropriate communications among the aforementioned components.
- the processor 382 may be a hardware device for executing software, particularly software stored in volatile and nonvolatile memory 384 .
- the processor 382 can be any custom made or commercially available processor, a central processing unit (CPU), an auxiliary processor among several processors associated with the computing device 104 , a semiconductor based microprocessor (in the form of a microchip or chip set), a macroprocessor, or generally any device for executing software instructions.
- CPU central processing unit
- auxiliary processor among several processors associated with the computing device 104
- semiconductor based microprocessor in the form of a microchip or chip set
- macroprocessor or generally any device for executing software instructions.
- the volatile and nonvolatile memory 384 can include any one or combination of volatile memory elements (e.g., random access memory (RAM, such as DRAM, SRAM, SDRAM, VRAM, etc.)) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CD-ROM, etc.). Moreover, the memory 384 may incorporate electronic, magnetic, optical, and/or other types of storage media. Note that the volatile and nonvolatile memory 384 can also have a distributed architecture, where various components are situated remotely from one another, but can be accessed by the processor 382 .
- the software in volatile and nonvolatile memory 384 may include one or more separate programs, each of which includes an ordered listing of executable instructions for implementing logical functions.
- the software in the volatile and nonvolatile memory 384 may include speech to text logic 388 , analytics logic 399 , as well as an operating system 386 , and a recording cache 397 .
- logic components 386 , 388 , and 390 are each illustrated in this nonlimiting example as a single piece of logic, as one of ordinary skill in the art will understand, these components can include one or more separate software, hardware, or firmware modules. Similarly, one or more of these logical components can be combined to provide the desired functionality.
- the operating system 386 may be configured to control the execution of other computer programs and may be configured to provide scheduling, input-output control, file and data management, memory management, and communication control and related services.
- a system component embodied as software may also be construed as a source program, executable program (object code), script, or any other entity comprising a set of instructions to be performed.
- the program is translated via a compiler, assembler, interpreter, or the like, which may or may not be included within the volatile and nonvolatile memory 384 , so as to operate properly in connection with the Operating System 386 .
- the Input/Output devices that may be coupled to system I/O Interface(s) 396 may include input devices, for example but not limited to, a keyboard, mouse, scanner, microphone, camera, proximity device, etc. Further, the Input/Output devices may also include output devices, for example but not limited to, a printer, display, etc. Finally, the Input/Output devices may further include devices that communicate both as inputs and outputs, for instance but not limited to, a modulator/demodulator (modem for accessing another device, system, or network), a radio frequency (RF) or other transceiver, a telephonic interface, a bridge, a router, etc.
- a modulator/demodulator modem for accessing another device, system, or network
- RF radio frequency
- network interface 388 which is coupled to local interface 392 can be configured to communication with a communications network, such as the network from FIGS. 2 and 3 . While this communication may be facilitated via a communications device, such as communications device 106 , this is not a requirement.
- the software in the volatile and nonvolatile memory 384 may further include a basic input output system (BIOS) (omitted for simplicity).
- BIOS is a set of software routines that initialize and test hardware at startup, start the Operating System 386 , and support the transfer of data among the hardware devices.
- the BIOS is stored in ROM so that the BIOS can be executed when the computing device 104 is activated.
- the processor 382 can be configured to execute software stored within the volatile and nonvolatile memory 384 , to communicate data to and from the volatile and nonvolatile memory 384 , and to generally control operations of the computing device 104 pursuant to the software.
- Software in memory, in whole or in part, is read by the processor 382 , perhaps buffered within the processor 382 , and then executed.
- other devices such as communications device 104 , computing device 108 , call center 106 , and/or other components can also include the components and/or functionality described in FIG. 3 .
- advanced data analytics component 218 can be configured with one or more of the components and/or logic described above with respect to analytics component 218 .
- analytics component 218 , communications device 104 , computing device 108 , and/or other components of call center 106 can include voice recognition logic, voice-to-text logic, text-to-voice logic, etc. (or any permutation thereof), as well as other components and/or logic for facilitating the functionality described herein. Additionally, in some exemplary embodiments, one or more of these components can include the functionality described with respect to analytics component 218 .
- FIG. 4 is an exemplary functional flow diagram, illustrating components that may be utilized for a phonetics based analytic infrastructure, as may be utilized in the network of FIG. 1 . As illustrated in the nonlimiting example of FIG. 4 , the components of FIG. 4 may be included as part of call center 106 . More specifically, speech recognition processing component 208 can receive an audio recording via extraction filtering component 206 and recorder subsystem component 204 . Speech processing framework 208 may receive this data in an ingestion component 404 and more specifically in a normalized audio component 406 . A normalized audio component 406 may be configured to facilitate the digitizing and or other processing of an audio recording.
- the normalized audio component 406 (which may be part of the phonetic engine) can send normalized audio to an ingest audio component 408 .
- the ingest audio component 408 may be configured to receive the normalized audio and one or more language models 410 for ingesting the communications data into a phonetic transcript.
- ingest audio component 408 can be configured to facilitate the creation of a phonetic transcript with one or more phonemes that occur in the communication.
- a representation of the one or more phonemes can include the designation of International Phonetic Alphabet (IPA) which may be utilized for computer use using the IS010646 standard (UNICODE).
- Ingest audio component 408 can then create the phonetic transcript 412 .
- the phonetic transcript 412 can then be sent to a search system 420 , which is part of a search component 416 .
- the search system can also receive vocabulary and rules as designated by an analyst, such as analyst 230 from FIG. 2 .
- Vocabulary and rules component 418 can designate one or more search trends and/or algorithms to perform in analyzing the phonetic transcript 412 .
- search terms 414 may be provided by analyst 230 to provide data associated with the quality of customer service provided to a caller.
- a user of communications device 104 may initiate communication with call center 106 .
- the user of communications device 104 and an agent 228 may engage in a communication that includes the phrase “brown fox.”
- the agent 228 may be required to save the term “brown fox.”
- the communication can be recorded and the audio recording 402 can be sent to the base phonetic engine.
- the phonetic engine can apply language models 410 to ingest the audio recording at ingest audio component 408 .
- the phonetic engine can thereby create a phonetic transcript 412 .
- the phonetic transcript can then be sent to a search component 416 , which includes a search system 420 .
- the search system 420 can utilize vocabulary and rules component 418 , as well as receive the search terms 414 .
- the search term “brown fox” can be a desired term to be found in a communication.
- the search system 420 can then search the phonetic transcript for the term “brown fox.”
- vocabulary and rules component 418 may be configured to provide a correlation between the search term 414 (which may be provided in English) and the phonetic representation of the desired search terms.
- a signal and/or scorecard can be provided to an analyst 230 to determine the quality of customer service provided by agent 228 . Additionally, some embodiments can be configured to provide information to analyst 230 in the event that the term “brown fox” does not appear in the communication. Similarly, other search terms and/or search criteria may be utilized to provide data to analyst 230 . Further description of phonetic speech to text conversion and analytics is disclosed in U.S. application Ser. No. ______, entitled “Speech Analysis Using Statistical Learning,” which is hereby incorporated by reference in its entirety.
- FIG. 5 is an exemplary functional block diagram illustrating speech recognition according to a text speech engine, similar to the diagram from FIG. 4 .
- an audio recording 502 may be provided to a text to speech engine 504 , where a normalized audio component 506 may receive an audio recording, normalize the audio recording, and provide the normalized audio recording to an ingest audio component 508 .
- Ingest audio component 508 can receive the normalized audio data, as well as text speech models 510 .
- Ingest audio component 508 may be configured to then convert the received audio recording into a textual transcript in a desired spoken language format (English, Spanish, French, etc.).
- the textual transcript 512 may then be provided to a search component 516 .
- Search system 512 (which may be part of search system 516 ), may receive the text transcript 512 , as well as search terms 514 . With search terms 514 and text transcript 512 , search component 516 may be configured to determine the quality of a communication between an agent and a customer.
- FIG. 5 may be included as part of call center 106 . More specifically, in at least one exemplary embodiment, one or more of the components of FIG. 5 may be included with speech processing framework 208 , advanced data analytics component 218 , and/or other components. Similarly, depending on the particular configuration, at least a portion of this logic may be provided with communications device 106 and/or computing device 108 .
- FIG. 6 is a network diagram illustrating a multi-tier speech recognition configuration, similar to the network configuration from FIG. 1 .
- communications device 104 may be coupled with local network 100 .
- a first tier recognition server 606 may be coupled to network 100 .
- a second tier speech recognition server 608 may be coupled to first tier recognition server 606 .
- first tier recognition server may be configured to receive audio data associated with a communication between a user of communications device 104 and an agent 228 associated with a call center 106 .
- First tier recognition server 606 may include a phonetic recognition server and/or a LVCSR server with a reduced accuracy rate. Regardless, first tier recognition server is configured to operate in an expedited mode, such that speech recognition and analytics of the received communication data is provided.
- first tier recognition server 606 is configured to provide speech to text conversion for received audio data.
- First tier recognition server 606 may be configured to convert speech to text via a phonetic engine (such as the phonetic engine described in FIG. 4 ) and/or via LVCSR (such as described in FIG. 5 ).
- second tier recognition server 608 may be configured to provide one or more speech to text analytics for the converted audio data. This may facilitate faster operation of the network in that first tier server 606 may be configured to, in at least one embodiment, process data faster than second tier server 608 . This may be accomplished via a faster computer architecture, utilization of faster speech to text recognition logic, and/or for other reasons.
- first tier server 606 is configured to determine the data to be sent to second tier server 608 . As a subset of data received at first tier server 606 is sent to second tier server 608 , the overall network performance may improve.
- first tier recognition server 606 is configured to provide a precursory speech to text conversion and/or analytics.
- first tier recognition server 606 can provide at least a portion of the communication data to second tier recognition server 608 .
- Second tier recognition server 608 may be configured to provide a more thorough analysis (and/or conversion) of the data.
- first tier server 606 may be configured to process at least a portion of the received data and send at least a portion of that data to second tier server 608 , network performance may improve.
- first tier recognition server 606 is illustrated as being directly coupled to second tier recognition server 608 , this is a nonlimiting example. More specifically, in at least one embodiment, first tier recognition server is coupled to network 100 and second tier recognition server is also coupled to network 100 . First tier recognition server 606 may be separately located from second tier recognition server and may facilitate communications with second tier recognition server 608 via network 100 . Additionally, while first tier recognition server 606 and second tier recognition server 608 are illustrated as separate components, this is also a nonlimiting example. In at least one embodiment, the functionality described with respect to first tier recognition server 606 and second tier recognition server 608 may be provided in a single component for providing the desired functionality.
- FIG. 7 is an exemplary network configuration illustrating a multi-tier speech recognition configuration with a plurality of servers, similar to the network configuration from FIG. 6 .
- communications device 104 is coupled to network 100 .
- first tier recognition servers 706 a , 706 b , and 706 c are also coupled to network 100 .
- first tier recognition servers 706 are coupled to first tier recognition servers 706 .
- First tier recognition server 706 may be configured to provide one or more speech recognition and/or analytics services.
- first tier recognition server 706 a may be configured to determine speaker identification associated with the communication.
- first tier server 706 b may be configured to provide speaker verification associated with the communication.
- First tier server 706 c may be configured to determine speaker emotion.
- second tier speech recognition service 708 a may be configured to exclusively serve first tier recognition server 706 a , however this is a nonlimiting example. More specifically, second tier speech recognition server 708 a may be configured as a speaker identification determination server to receive data from first recognition server 706 a .
- audio data may be sent to first tier speech recognition server 706 a , which may be a phonetic speech recognition server.
- First tier speech recognition server 706 a may be configured to determine at least one characteristic associated with the audio data to determine whether speaker identification may be determined. If a determination is made that speaker identification can be determined, first tier speech recognition service 706 a may send at least a portion of the received audio data (which may be converted into a phonetic transcript and/or other form) to a second tier recognition server 708 a .
- Second tier speech recognition server 708 a may be configured to fully analyze the received data to determine the identification of the speaker.
- second tier speech recognition server 708 a is a dedicated server for first tier speech recognition server 706 a
- second tier recognition servers may serve one or more of the first tier speech recognition servers 706 .
- first tier recognition server 706 may be configured to provide the initial speech recognition functionality while second tier speech recognition server 708 may be configured to provide more specific services.
- first tier speech recognition servers 706 a , 706 b , 706 c may be configured to provide a speech to text conversion associated with received audio data.
- first tier speech recognition servers 706 a , 706 b , and 706 c can make a determination as to the desired analytics for the associated communication.
- first tier speech recognition server 706 can send the phonetic data to a second tier speech recognition server 708 associated with the desired analytic.
- second tier speech recognition server 708 a is a speaker identification server
- one or more of the first tier recognition servers 706 a , 706 b , 706 c can send data to second tier speech recognition server 708 a upon determination that a speaker identification is required.
- speech recognition servers 706 may be configured to send communication data to second tier speech recognition server 708 b .
- Other configurations are also included.
- FIG. 8 is an exemplary network configuration illustrating a three-tier speech recognition configuration, similar to the network configuration from FIG. 7 .
- communications device 104 is connected to network 100 .
- first tier speech recognition server 806 is Also connected to network 100 .
- first tier speech recognition server 806 may be configured to provide expedited speech recognition services via a phonetics speech to text configuration and/or an LVCSR functionality with reduced accuracy such that processing of the audio data may be completed in an expedited manner.
- FIG. 8 Also included in the nonlimiting example of FIG. 8 is a second tier speech recognition server 808 .
- Second tier speech recognition server 808 may be configured to provide speech recognition services at higher accuracy than first tier speech recognition server 806 .
- FIG. 8 also includes a third tier recognition server 810 , which may be configured to further speech recognition analytics associated with the communication.
- first tier speech recognition server 806 may be configured to receive raw data associated with a communication. First tier speech recognition server 806 may then perform expedited speech recognition services on the received data. Second tier speech recognition server 808 may include more thorough speech recognition functionality which may be slower in operation than first tier speech recognition server 806 , however second tier server 808 may provide greater accuracy related to received data. Additionally, second tier speech recognition server 808 may make a determination whether a third tier speech recognition server 810 may be utilized.
- Third tier speech recognition server 810 may be configured to provide services different than that of second tier speech recognition server 808 .
- second tier speech recognition server 808 may be configured to determined speaker confidence associated with received audio data
- a third tier speech recognition server may be configured to determine speaker emotion associated with the received audio.
- utilization of second tier speech recognition server 808 and third tier speech recognition server 810 (as well as first tier speech recognition server 806 ) may be utilized.
- FIG. 9A is a flowchart illustrating an exemplary embodiment of a process that may be utilized to perform speech recognition functionality in a multi-tier speech recognition environment, such as the environment from FIG. 6 .
- call center 106 may receive raw audio data associated with a communication on a first tier speech recognition server (block 932 ).
- Call server 106 may then determine recognition criteria (block 934 ). More specifically, call center 106 may be configured to determine the type of speech recognition analysis that may be performed on the received communication data.
- the received communication call center 106 may determine that the first tier speech recognition server determines speaker identification associated with communication. This indication may be received from an agent 228 , analyst 230 , and/or supervisor 232 , depending upon the embodiment.
- Call center 106 may then convert the received audio into a textual transcript (e.g., a phonetic transcript and/or a spoken language transcript and/or other type of transcript), as illustrated in block 936 .
- Call center 106 may then determine whether the audio potentially include the recognition criteria (block 938 ). If the received audio data does not include the recognition criteria, the process may end. If, however, first tier speech recognition server determines that the audio potentially include the recognition criteria, first tier speech recognition server can send at least a portion of the audio (which may be converted to a phonetic and/or other transcript) to second tier speech recognition server (block 938 ). The flowchart then proceeds to jump block 940 , which is continued in FIG. 9B .
- a textual transcript e.g., a phonetic transcript and/or a spoken language transcript and/or other type of transcript
- FIG. 9B is a continuation from the flowchart from FIG. 9A .
- a flowchart begins from a jump block 942 , which is continued from FIG. 9A .
- the call center 106 may then receive data from the first tier speech recognition server at a second tier speech recognition server (block 944 ).
- the second tier speech recognition server may then perform full recognition analytics according to the recognition criteria (block 946 ).
- second tier speech recognition server may provide a more detailed speech recognition analysis of the audio data received.
- some embodiments may be configured to provide a specific speech recognition analysis task such as speaker identification, speaker verification, speaker emotion, speaker confidence, and/or other types of analysis.
- FIG. 10A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing speech recognition in a first tier upon a determination that utilization of a second speech recognition tier is desired, similar to the flowchart from FIGS. 9A and 9B .
- call center 106 may receive raw audio data at a first tier speech recognition server (block 1032 ).
- the first tier recognition server may then determine recognition criteria associated with the received audio data (block 1034 ).
- First tier speech recognition server can then begin speech to text conversion (block 1036 ).
- first tier speech recognition server determines that the received valued data potentially include recognition criteria (block 1038 )
- first tier speech recognition server can stop the conversion and send the audio data to the second tier speech recognition server (block 1040 ).
- the flowchart can then proceed to jump block 1042 .
- FIG. 10B is a continuation of the flowchart from FIG. 10A .
- the second tier speech recognition server can receive data from the first tier speech recognition server (block 1046 ). Second tier speech recognition server can then perform full recognition according to the recognition criteria (block 1048 ).
- first tier speech recognition server can send at least a portion of the data to the second tier speech recognition server. As such, full analysis of the received audio may be expedited.
- FIG. 11A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing a user option for providing an agent an option to send audio data to a second tier speech recognition system, similar to the flowchart from FIGS. 10A and 10B .
- call center 106 can receive raw audio data at a first tier speech recognition server (block 1132 ).
- the first tier speech recognition server can then determine the recognition criteria associated with the received audio data (block 1134 ).
- the first tier speech recognition server can then convert the received audio into a textual format, as described above (block 1136 ).
- the first tier speech recognition server can then send at a least a portion of the conversion results to an analyst 230 (block 1138 ).
- the analyst 230 can then indicate whether to send the audio data to a second tier speech recognition server (block 1140 ). If the analyst 230 determines that the data need not be sent to a second tier speech recognition server, the process may end. If, however, the analyst 230 determines that further speech recognition analysis may be determined on the received audio data, the first tier speech recognition server can send at least a portion of the received audio data to a second tier speech recognition server (block 1142 ). Flowchart can then proceed to jump block 1143 .
- FIG. 11B is a continuation of the flowchart from FIG. 11A .
- the second tier speech recognition server can receive at least a portion of the audio data from the first speech recognition server (block 1146 ).
- the second tier speech recognition server can then perform full recognition of the received audio data according to the recognition criteria (block 1148 ).
- second tier speech recognition server may also be configured to perform specific speech recognition analytics, such as speaker identification and/or other specific speech analytics.
- FIG. 12A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for determining from a plurality of second tier speech recognition servers, which server tier direct audio data, similar to the flowchart from FIGS. 11A and 11B .
- call center 106 can receive raw audio data at a first tier speech recognition server (block 1232 ). The call center 106 can then determine recognition criteria associated with the received audio data (block 1234 ). The first tier speech recognition server can then begin conversion of the received audio data into a textual format using phonetic speech to text conversion and/or other expedited speech to text conversion techniques (block 1236 ). Call center 106 may then determine whether the audio potentially include the recognition criteria (block 1238 ). If a determination is made that the received audio does not include the recognition criteria, the flowchart can end. If, however, a determination is made that the audio potentially include the recognition criteria, the flowchart can proceed to jump block 1240 .
- FIG. 12B is a continuation of the flowchart from FIG. 12A .
- call center 106 can determine which second tier speech recognition server to be utilized for the received audio data (block 1244 ).
- Call center 106 can then receive data from the first tier speech recognition server at the determined second tier speech recognition server (block 1246 ).
- the second tier speech recognition server can then perform full recognition of the audio data according to the recognition criteria (block 1248 ).
- speech analytics i.e., the analysis of recorded speech or real-time speech
- speech analytics can be used to perform a variety of functions, such as automated call evaluation, call scoring, quality monitoring, quality assessment and compliance/adherence.
- speech analytics can be used to compare a recorded interaction to a script (e.g., a script that the agent was to use during the interaction).
- speech analytics can be used to measure how well agents adhere to scripts, identify which agents are “good” sales people and which ones need additional training. As such, speech analytics can be used to find agents who do not adhere to scripts.
- speech analytics can measure script effectiveness, identify which scripts are effective and which are not, and find, for example, the section of a script that displeases or upsets customers (e.g., based on emotion detection).
- compliance with various policies can be determined. Such may be in the case of, for example, the collections industry where it is a highly regulated business and agents must abide by many rules.
- the speech analytics of the present disclosure may identify when agents are not adhering to their scripts and guidelines. This can improve collection effectiveness and reduce corporate liability and risk.
- various types of recording components can be used to facilitate speech analytics.
- such recording components can perform one or more various functions such as receiving, capturing, intercepting and tapping of data. This can involve the use of active and/or passive recording techniques, as well as the recording of voice and/or screen data.
- speech analytics can be used in conjunction with such screen data (e.g., screen data captured from an agent's workstation/PC) for evaluation, scoring, analysis, adherence and compliance purposes, for example.
- screen data e.g., screen data captured from an agent's workstation/PC
- speech analytics can be used in conjunction with such screen data (e.g., screen data captured from an agent's workstation/PC) for evaluation, scoring, analysis, adherence and compliance purposes, for example.
- Such integrated functionalities improve the effectiveness and efficiency of, for example, quality assurance programs.
- the integrated function can help companies to locate appropriate calls (and related screen interactions) for quality monitoring and evaluation. This type of “precision” monitoring improves the effectiveness and productivity of quality assurance programs.
- speech analytics can be used independently and/or in combination with other techniques for performing fraud detection. Specifically, some embodiments can involve identification of a speaker (e.g., a customer) and correlating this identification with other information to determine whether a fraudulent claim for example is being made. If such potential fraud is identified, some embodiments can provide an alert.
- the speech analytics of the present disclosure may identify the emotions of callers. The identified emotions can be used in conjunction with identifying specific concepts to help companies spot either agents or callers/customers who are involved in fraudulent activities.
- At least one embodiment of an integrated workforce optimization platform integrates: (1) Quality Monitoring/Call Recording—voice of the customer; the complete customer experience across multimedia touch points; (2) Workforce Management—strategic forecasting and scheduling that drives efficiency and adherence, aids in planning, and helps facilitate optimum staffing and service levels; (3) Performance Management—key performance indicators (KPIs) and scorecards that analyze and help identify synergies, opportunities and improvement areas; (4) e-Learning—training, new information and protocol disseminated to staff, leveraging best practice customer interactions and delivering learning to support development; and/or (5) Analytics—deliver insights from customer interactions to drive business performance.
- KPIs key performance indicators
- e-Learning training, new information and protocol disseminated to staff, leveraging best practice customer interactions and delivering learning to support development
- Analytics delivery insights from customer interactions to drive business performance.
- the integrated workforce optimization process and system can include planning and establishing goals—from both an enterprise and center perspective—to ensure alignment and objectives that complement and support one another.
- planning may be complemented with forecasting and scheduling of the workforce to ensure optimum service levels.
- Recording and measuring performance may also be utilized, leveraging quality monitoring/call recording to assess service quality and the customer experience.
- the embodiments disclosed herein can be implemented in hardware, software, firmware, or a combination thereof. At least one embodiment, disclosed herein is implemented in software and/or firmware that is stored in a memory and that is executed by a suitable instruction execution system. If implemented in hardware, as in an alternative embodiment disclosed herein can be implemented with any or a combination of the following technologies: a discrete logic circuit) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), etc.
- ASIC application specific integrated circuit
- PGA programmable gate array
- FPGA field programmable gate array
- each block can be interpreted to represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the blocks may occur out of the order and/or not at all. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- any of the programs listed herein can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions.
- a “computer-readable medium” can be any means that can contain, store, communicate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- the computer readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device.
- the computer-readable medium could include an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM or Flash memory) (electronic), an optical fiber (optical), and a portable compact disc read-only memory (CDROM) (optical).
- the scope of the certain embodiments of this disclosure can include embodying the functionality described in logic embodied in hardware or software-configured mediums.
- conditional language such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more particular embodiments or that one or more particular embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Telephonic Communication Services (AREA)
Abstract
Description
- This application is a continuation application of and claims priority to U.S. patent application Ser. No. 11/540,322, titled MULTI-PASS SPEECH ANALYTICS, filed on Sep. 29, 2006, and which is hereby incorporated by reference in its entirety. No new matter has been added.
- As communication technologies have improved, businesses and individuals have desired greater functionality in their communication networks. As a nonlimiting example, many businesses have created call center infrastructures in which a customer or other user can call to receive information related to the business. As customers call into the call center, the customer may be connected with a customer service representative to provide the desired information. Depending on the time of call, the subject matter of the call, and/or other data, the customer may be connected with different customer service representatives. As such, depending on these and/or other factors, the customer may be provided with varying levels of customer service with respect to the interaction with the customer service representative. Because most businesses desire to provide the highest possible quality of customer service, many businesses have turned to recording the communication between the customer and the customer service representative. While recording this data has proven beneficial in many cases, many businesses receive call volumes that inhibit the business from reviewing all of the call data received.
- As such, many businesses have turned to speech recognition technology to capture the recorded communication data and thereby provide a textual document for review of the communication. While textual documentation of a communication has also proven beneficial, similar issues may exist in that the sheer amount of data may be such that review of the data is impractical.
- To combat this problem, a number of businesses have also implemented analytics technologies to analyze the speech-recognized communications. One such technology that has emerged includes large vocabulary continuous speech recognition (LVCSR). LVCSR technologies often convert received audio from the communications into an English translation of the communication in a textual document. From the textual document, analytics may be provided to determine various data related to the communication. Additionally, phonetic speech recognition may be utilized for capturing the communication data.
- While these and technologies may provide a mechanism for capturing communication data, oftentimes, the shear amount of data for processing may consume extensive hardware resources. As such, a solution to increase speed and/or reduce resource consumption is desired.
- Included are embodiments for multi-pass analytics. At least one embodiment of a method includes receiving audio data associated with first communications desired to be analyzed for an event, performing a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and performing a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- Also included are embodiments of a system for multi-pass analytics. At least one embodiment of a system includes an interface configured to receive audio data associated with first communications desired to be analyzed for an event, and a processor coupled to the interface and configured to perform a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and perform a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- Also included are embodiments of a computer readable medium for multi-pass analytics. At least one embodiment includes a computer readable medium having stored thereon instructions that, when executed by a processor, direct the processor to receive audio data associated with first communications desired to be analyzed for an event, perform a phonetics analytics process on the audio data associated with the first communications to determine second communications associated with the event, and perform a large vocabulary continuous speech recognition (LVCSR) analytics process on audio data associated with the second communications to determine third communications associated with the event.
- Other systems, methods, features, and advantages of this disclosure will be or become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features, and advantages be included within this description, be within the scope of the present disclosure.
- Many aspects of the disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views. While several embodiments are described in connection with these drawings, there is no intent to limit the disclosure to the embodiment or embodiments disclosed herein. On the contrary, the intent is to cover all alternatives, modifications, and equivalents.
-
FIG. 1 is an exemplary embodiment of a system in which a user may be provided with customer service. -
FIG. 2 is an exemplary embodiment of a recording network configuration, in which recording and analytics may be provided, similar to the network configuration fromFIG. 1 . -
FIG. 3 is a functional block diagram illustrating an exemplary embodiment of an analytic component, similar to the analytic component fromFIG. 2 . -
FIG. 4 is an exemplary embodiment of a functional flow diagram, illustrating components that may be utilized for a phonetics based analytic infrastructure, as may be utilized in the network ofFIG. 1 . -
FIG. 5 is an exemplary embodiment of a functional block diagram illustrating speech recognition according to a text speech engine, similar to the diagram fromFIG. 4 . -
FIG. 6 is a network diagram illustrating an exemplary embodiment of a multi-tier speech recognition configuration, similar to the network configuration fromFIG. 1 . -
FIG. 7 is a network configuration illustrating an exemplary embodiment of a multi-tier speech recognition configuration with a plurality of servers, similar to the network configuration fromFIG. 6 . -
FIG. 8 is a network configuration illustrating an exemplary embodiment of a three-tier speech recognition configuration, similar to the network configuration fromFIG. 7 . -
FIG. 9A is a flowchart illustrating the exemplary embodiment of a process that may be utilized to perform speech recognition functionality in a multi-tier speech recognition environment, such as the environment fromFIG. 6 . -
FIG. 9B is a continuation from the flowchart fromFIG. 9A . -
FIG. 10A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing speech recognition in a first tier upon a determination that utilization of a second speech recognition tier is desired, similar to the flowchart fromFIGS. 9A and 9B . -
FIG. 10B is a continuation of the flowchart fromFIG. 10A . -
FIG. 11A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing a user option for providing an agent an option to send audio data to a second tier speech recognition system, similar to the flowchart fromFIGS. 10A and 10B . -
FIG. 11B is a continuation of the flowchart fromFIG. 11A . -
FIG. 12A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for determining from a plurality of second tier speech recognition servers, which server tier direct audio data, similar to the flowchart fromFIGS. 11A and 11B . -
FIG. 12B is a continuation of the flowchart fromFIG. 12A . - Included are embodiments for increasing the speed of speech to text conversion and related analytics. More specifically, in at least one embodiment, first tier speech to text analytics and second tier speech to text analytics are used. In other embodiments, a first tier may be configured for speech to text conversion and a second tier may be configured for speech to text analytics. Other embodiments are also included, as discussed with reference to the drawings.
-
FIG. 1 is a nonlimiting example of a system in which a user may be provided with customer service. As illustrated in this nonlimiting example, a caller can initiate a communication request from acommunications device 104. The communications request can be sent tocall center 106 vianetwork 100.Network 100 may include a Public Switched Telephone Network (PSTN), an Internet Protocol (IP) network, a cellular network, an Integrated Services Digital Network (ISDN), a Time Division Multiplexing (TDM) network, and/or other networks. Upon connecting withcall center 106, a user atcommunications device 104 may facilitate a communication with a customer service representative. Additionally, during the communication, the communication can be recorded, such that a business associated withcall center 106 can determine the quality of customer service provided to the user ofcommunications device 104. - While in some configurations, an audio recording can be provided to an analyst to determine the quality of customer service, some embodiments may include a text to voice conversion of the communication. Large Vocabulary Continuous Speech Recognition (LVCSR) may be utilized to create an English translated textual document associated with the communication. While an LVCSR speech recognized textual document may provide enhanced searching capabilities related to the communication, depending on an accuracy threshold, LVCSR technologies may be slow in execution. Similarly, in many phonetic technologies for speech recognition, processing of search functions associated with the communication may be slow.
- Additionally, while a user can send a communication request via
communication device 104, some embodiments may provide that a user utilizingcomputing device 108 may initiate a communication tocall center 106 vianetwork 100. In such configurations, a user may utilize a soft phone and/or other communications logic provided for initiating and facilitating a communication. - One souls also note that a call center can include, but is not limited to, outsourced contact centers, outsourced customer relationship management, customer relationship management, voice of the customer, customer interaction, contact center, multi-media contact center, remote office, distributed enterprise, work-at-home agents, remote agents, branch office, back office, performance optimization, workforce optimization, hosted contact centers, and speech analytics, for example.
-
FIG. 2 is a nonlimiting example of a system for providing recording and analytics, similar to the network configuration fromFIG. 1 . As illustrated in the nonlimiting example ofFIG. 2 ,communications device 104 can facilitate communications withcall center 106 vianetwork 100.Call center 106 can include one or more elements, such as arecorder subsystem component 204, anextraction filtering component 206, and a speechprocessing firmware component 208. Also included incall center 106 is an advanced data analytics (pattern recognition)component 218 and anatlas component 210. Included in theatlas component 210 is aspeech package 212, adesktop event 214, and aninteractions package 216. -
Call center 106 may also include ananalytic scorecard 220, a quality management (QM)evaluations component 222, andenterprise reporting component 224, and a speech andreplay component 226. Anagent 228 can utilize one or more of the components ofcall center 106 to facilitate a communication with a caller oncommunications device 104. Similarly, ananalyst 230 can utilize one or more components ofcall center 106 to analyze the quality of the communications between theagent 228 and the caller associated withcommunications device 104. Asupervisor 232 may also have access to components ofcall center 106 to oversee theagent 228 and/or theanalyst 230 and their interactions with a caller oncommunications device 104. - Additionally, a
recognition engine cluster 202 may be coupled tocall center 106 directly and/or vianetwork 100.Recognition engine cluster 202 may include one or more servers that may provide speech recognition functionality to callcenter 106. In operation, a communication between a caller oncommunications device 104 and anagent 228, vianetwork 100, may first be received by arecorder subsystem component 204.Recorder subsystem component 204 may record the communications in an audio format. The recorder audio may then be sent to anextraction filtering component 206 which may be configured to extract the dialogue (e.g., remove noise and other unwanted sounds) from the recording. The recorded communication can then be sent to a speech-processingframework component 208 for converting the recorded audio communication into a textual format. Conversion of the audio into a textual format may be facilitated by arecognition engine cluster 202, however this is not a requirement. Regardless, conversion from the audio format to a textual format may be facilitated via LVCSR speech recognition technologies and/or phonetic speech recognition technologies, as discussed in more detail below. - Upon conversion from audio to a textual format, data related to the communication may be provided to advanced data analytics (pattern recognition)
component 218. Advanceddata analytics component 218 may be configured to provide analysis associated with the speech to text converted communication to determine the quality of customer service provided to the caller ofcommunications device 104. Advanceddata analytics component 218 may utilizeatlas component 210 for facilitation of this analysis. More specifically,atlas component 210 may include aspeech package component 212 that may be configured to analyze various patterns in the speech of the caller ofcommunications device 104. Similarly,desktop event component 214 may be configured to analyze one or more actions that the user of communications device takes on theircommunications device 104. More specifically, anetwork 100 may facilitate communications in an IP network. As such,communications device 104 may facilitate both audio and/or data communications that may include audio, video, images, and/or other data. Additionally, advanceddata analytics component 218 may utilize anactions package 216 to determine various components of the interaction betweenagent 228 and the caller ofcommunications device 104. Advanceddata analytics component 218 may then make a determination based on predetermined criteria of the quality of call service provided byagent 220. - Advanced
data analytics component 218 may then facilitate creation of ananalytic scorecard 220 and provideenterprise reporting 224. Additionally, call center may providequality management evaluations 222, as well as speech andreplay communications 226. This data may be viewed by anagent 228, ananalyst 230, and/or asupervisor 232. Additionally, as discussed in more detail below, ananalyst 230 may further analyze the data to provide a basis for advanceddata analytics component 218 to determine the quality of customer service. -
FIG. 3 is a functional block diagram illustrating an analytic component, similar to the analytic component fromFIG. 2 . Although a wire-line communications device is illustrated, this discussion can be applied to any device configured for receiving and/or sending data. As illustrated inFIG. 3 , in terms of hardware architecture, advanceddata analytics component 218 includes aprocessor 382, volatile andnonvolatile memory 384, adisplay interface 394,data storage 395, and one or more input and/or output (I/O) device interface(s) 396 that are communicatively coupled via alocal interface 392. Thelocal interface 392 can include, for example but not limited to, one or more buses and/or other wired or wireless connections. Thelocal interface 392 may have additional elements, which are omitted for simplicity, such as controllers, buffers (caches), drivers, repeaters, and receivers to enable communications. Further, the local interface may include address, control, and/or data connections to enable appropriate communications among the aforementioned components. Theprocessor 382 may be a hardware device for executing software, particularly software stored in volatile andnonvolatile memory 384. - The
processor 382 can be any custom made or commercially available processor, a central processing unit (CPU), an auxiliary processor among several processors associated with thecomputing device 104, a semiconductor based microprocessor (in the form of a microchip or chip set), a macroprocessor, or generally any device for executing software instructions. - The volatile and
nonvolatile memory 384 can include any one or combination of volatile memory elements (e.g., random access memory (RAM, such as DRAM, SRAM, SDRAM, VRAM, etc.)) and nonvolatile memory elements (e.g., ROM, hard drive, tape, CD-ROM, etc.). Moreover, thememory 384 may incorporate electronic, magnetic, optical, and/or other types of storage media. Note that the volatile andnonvolatile memory 384 can also have a distributed architecture, where various components are situated remotely from one another, but can be accessed by theprocessor 382. - The software in volatile and
nonvolatile memory 384 may include one or more separate programs, each of which includes an ordered listing of executable instructions for implementing logical functions. In the example ofFIG. 3 , the software in the volatile andnonvolatile memory 384 may include speech to textlogic 388,analytics logic 399, as well as anoperating system 386, and a recording cache 397. Additionally, whilelogic components operating system 386 may be configured to control the execution of other computer programs and may be configured to provide scheduling, input-output control, file and data management, memory management, and communication control and related services. - A system component embodied as software may also be construed as a source program, executable program (object code), script, or any other entity comprising a set of instructions to be performed. When constructed as a source program, the program is translated via a compiler, assembler, interpreter, or the like, which may or may not be included within the volatile and
nonvolatile memory 384, so as to operate properly in connection with theOperating System 386. - The Input/Output devices that may be coupled to system I/O Interface(s) 396 may include input devices, for example but not limited to, a keyboard, mouse, scanner, microphone, camera, proximity device, etc. Further, the Input/Output devices may also include output devices, for example but not limited to, a printer, display, etc. Finally, the Input/Output devices may further include devices that communicate both as inputs and outputs, for instance but not limited to, a modulator/demodulator (modem for accessing another device, system, or network), a radio frequency (RF) or other transceiver, a telephonic interface, a bridge, a router, etc. Similarly,
network interface 388, which is coupled tolocal interface 392 can be configured to communication with a communications network, such as the network fromFIGS. 2 and 3 . While this communication may be facilitated via a communications device, such ascommunications device 106, this is not a requirement. - If the
computing device 104 is a personal computer, workstation, or the like, the software in the volatile andnonvolatile memory 384 may further include a basic input output system (BIOS) (omitted for simplicity). The BIOS is a set of software routines that initialize and test hardware at startup, start theOperating System 386, and support the transfer of data among the hardware devices. The BIOS is stored in ROM so that the BIOS can be executed when thecomputing device 104 is activated. - When the
computing device 104 is in operation, theprocessor 382 can be configured to execute software stored within the volatile andnonvolatile memory 384, to communicate data to and from the volatile andnonvolatile memory 384, and to generally control operations of thecomputing device 104 pursuant to the software. Software in memory, in whole or in part, is read by theprocessor 382, perhaps buffered within theprocessor 382, and then executed. Additionally, one should note that while the above description is directed to a advanceddata analytics component 218, other devices (such ascommunications device 104,computing device 108,call center 106, and/or other components) can also include the components and/or functionality described inFIG. 3 . - One should also note that advanced
data analytics component 218 can be configured with one or more of the components and/or logic described above with respect toanalytics component 218. Additionally,analytics component 218,communications device 104,computing device 108, and/or other components ofcall center 106 can include voice recognition logic, voice-to-text logic, text-to-voice logic, etc. (or any permutation thereof), as well as other components and/or logic for facilitating the functionality described herein. Additionally, in some exemplary embodiments, one or more of these components can include the functionality described with respect toanalytics component 218. -
FIG. 4 is an exemplary functional flow diagram, illustrating components that may be utilized for a phonetics based analytic infrastructure, as may be utilized in the network ofFIG. 1 . As illustrated in the nonlimiting example ofFIG. 4 , the components ofFIG. 4 may be included as part ofcall center 106. More specifically, speechrecognition processing component 208 can receive an audio recording viaextraction filtering component 206 andrecorder subsystem component 204.Speech processing framework 208 may receive this data in aningestion component 404 and more specifically in a normalizedaudio component 406. A normalizedaudio component 406 may be configured to facilitate the digitizing and or other processing of an audio recording. Additionally, the normalized audio component 406 (which may be part of the phonetic engine) can send normalized audio to an ingestaudio component 408. The ingestaudio component 408 may be configured to receive the normalized audio and one ormore language models 410 for ingesting the communications data into a phonetic transcript. - More specifically, ingest
audio component 408 can be configured to facilitate the creation of a phonetic transcript with one or more phonemes that occur in the communication. One embodiment of a representation of the one or more phonemes can include the designation of International Phonetic Alphabet (IPA) which may be utilized for computer use using the IS010646 standard (UNICODE). Ingestaudio component 408 can then create thephonetic transcript 412. - The
phonetic transcript 412 can then be sent to asearch system 420, which is part of asearch component 416. The search system can also receive vocabulary and rules as designated by an analyst, such asanalyst 230 fromFIG. 2 . Vocabulary and rules component 418 can designate one or more search trends and/or algorithms to perform in analyzing thephonetic transcript 412. Additionally, search terms 414 may be provided byanalyst 230 to provide data associated with the quality of customer service provided to a caller. - As a nonlimiting example, referring to
FIGS. 2 and 4 , a user ofcommunications device 104 may initiate communication withcall center 106. The user ofcommunications device 104 and anagent 228 may engage in a communication that includes the phrase “brown fox.” In this communication, theagent 228 may be required to save the term “brown fox.” As such, the communication can be recorded and theaudio recording 402 can be sent to the base phonetic engine. The phonetic engine can applylanguage models 410 to ingest the audio recording at ingestaudio component 408. The phonetic engine can thereby create aphonetic transcript 412. - The phonetic transcript can then be sent to a
search component 416, which includes asearch system 420. Thesearch system 420 can utilize vocabulary and rules component 418, as well as receive the search terms 414. As indicated above, the search term “brown fox” can be a desired term to be found in a communication. Thesearch system 420 can then search the phonetic transcript for the term “brown fox.” As the phonetic transcript may not include an English translation of the audio recording, vocabulary and rules component 418 may be configured to provide a correlation between the search term 414 (which may be provided in English) and the phonetic representation of the desired search terms. - If the term “brown fox” appears in the
phonetic transcript 412, a signal and/or scorecard can be provided to ananalyst 230 to determine the quality of customer service provided byagent 228. Additionally, some embodiments can be configured to provide information toanalyst 230 in the event that the term “brown fox” does not appear in the communication. Similarly, other search terms and/or search criteria may be utilized to provide data toanalyst 230. Further description of phonetic speech to text conversion and analytics is disclosed in U.S. application Ser. No. ______, entitled “Speech Analysis Using Statistical Learning,” which is hereby incorporated by reference in its entirety. -
FIG. 5 is an exemplary functional block diagram illustrating speech recognition according to a text speech engine, similar to the diagram fromFIG. 4 . As illustrated in the nonlimiting example ofFIG. 5 , anaudio recording 502 may be provided to a text tospeech engine 504, where a normalizedaudio component 506 may receive an audio recording, normalize the audio recording, and provide the normalized audio recording to an ingestaudio component 508. Ingestaudio component 508 can receive the normalized audio data, as well astext speech models 510. Ingestaudio component 508 may be configured to then convert the received audio recording into a textual transcript in a desired spoken language format (English, Spanish, French, etc.). Thetextual transcript 512 may then be provided to asearch component 516. Search system 512 (which may be part of search system 516), may receive thetext transcript 512, as well assearch terms 514. Withsearch terms 514 andtext transcript 512,search component 516 may be configured to determine the quality of a communication between an agent and a customer. - One should note that the nonlimiting example of
FIG. 5 may be included as part ofcall center 106. More specifically, in at least one exemplary embodiment, one or more of the components ofFIG. 5 may be included withspeech processing framework 208, advanceddata analytics component 218, and/or other components. Similarly, depending on the particular configuration, at least a portion of this logic may be provided withcommunications device 106 and/orcomputing device 108. -
FIG. 6 is a network diagram illustrating a multi-tier speech recognition configuration, similar to the network configuration fromFIG. 1 . As illustrated in the nonlimiting example ofFIG. 6 ,communications device 104 may be coupled withlocal network 100. Additionally, a firsttier recognition server 606 may be coupled tonetwork 100. Additionally, a second tierspeech recognition server 608 may be coupled to firsttier recognition server 606. In operation, first tier recognition server may be configured to receive audio data associated with a communication between a user ofcommunications device 104 and anagent 228 associated with acall center 106. Firsttier recognition server 606 may include a phonetic recognition server and/or a LVCSR server with a reduced accuracy rate. Regardless, first tier recognition server is configured to operate in an expedited mode, such that speech recognition and analytics of the received communication data is provided. - In at least one embodiment, the system of
FIG. 6 may be configured such that firsttier recognition server 606 is configured to provide speech to text conversion for received audio data. Firsttier recognition server 606 may be configured to convert speech to text via a phonetic engine (such as the phonetic engine described inFIG. 4 ) and/or via LVCSR (such as described inFIG. 5 ). In such a configuration, secondtier recognition server 608 may be configured to provide one or more speech to text analytics for the converted audio data. This may facilitate faster operation of the network in thatfirst tier server 606 may be configured to, in at least one embodiment, process data faster thansecond tier server 608. This may be accomplished via a faster computer architecture, utilization of faster speech to text recognition logic, and/or for other reasons. Similarly, some embodiments may be configured such thatfirst tier server 606 is configured to determine the data to be sent tosecond tier server 608. As a subset of data received atfirst tier server 606 is sent tosecond tier server 608, the overall network performance may improve. - Similarly, some embodiments may be configured such that first
tier recognition server 606 is configured to provide a precursory speech to text conversion and/or analytics. Upon recognition of a desired search term associated with the communication, firsttier recognition server 606 can provide at least a portion of the communication data to secondtier recognition server 608. Secondtier recognition server 608 may be configured to provide a more thorough analysis (and/or conversion) of the data. Asfirst tier server 606 may be configured to process at least a portion of the received data and send at least a portion of that data tosecond tier server 608, network performance may improve. - While first
tier recognition server 606 is illustrated as being directly coupled to secondtier recognition server 608, this is a nonlimiting example. More specifically, in at least one embodiment, first tier recognition server is coupled tonetwork 100 and second tier recognition server is also coupled tonetwork 100. Firsttier recognition server 606 may be separately located from second tier recognition server and may facilitate communications with secondtier recognition server 608 vianetwork 100. Additionally, while firsttier recognition server 606 and secondtier recognition server 608 are illustrated as separate components, this is also a nonlimiting example. In at least one embodiment, the functionality described with respect to firsttier recognition server 606 and secondtier recognition server 608 may be provided in a single component for providing the desired functionality. -
FIG. 7 is an exemplary network configuration illustrating a multi-tier speech recognition configuration with a plurality of servers, similar to the network configuration fromFIG. 6 . As illustrated in the nonlimiting example ofFIG. 7 ,communications device 104 is coupled tonetwork 100. Also coupled tonetwork 100 are firsttier recognition servers tier recognition server - First tier recognition server 706 may be configured to provide one or more speech recognition and/or analytics services. As a nonlimiting example, first
tier recognition server 706 a may be configured to determine speaker identification associated with the communication. Similarly,first tier server 706 b may be configured to provide speaker verification associated with the communication.First tier server 706 c may be configured to determine speaker emotion. Similarly, second tierspeech recognition service 708 a may be configured to exclusively serve firsttier recognition server 706 a, however this is a nonlimiting example. More specifically, second tierspeech recognition server 708 a may be configured as a speaker identification determination server to receive data fromfirst recognition server 706 a. In operation, audio data may be sent to first tierspeech recognition server 706 a, which may be a phonetic speech recognition server. First tierspeech recognition server 706 a may be configured to determine at least one characteristic associated with the audio data to determine whether speaker identification may be determined. If a determination is made that speaker identification can be determined, first tierspeech recognition service 706 a may send at least a portion of the received audio data (which may be converted into a phonetic transcript and/or other form) to a secondtier recognition server 708 a. Second tierspeech recognition server 708 a may be configured to fully analyze the received data to determine the identification of the speaker. - While the nonlimiting example discussed above indicates that the second tier
speech recognition server 708 a is a dedicated server for first tierspeech recognition server 706 a, this is a nonlimiting example. More specifically, in at least one nonlimiting example, second tier recognition servers may serve one or more of the first tier speech recognition servers 706. Similarly, some embodiments can be configured such that first tier recognition server 706 may be configured to provide the initial speech recognition functionality while second tier speech recognition server 708 may be configured to provide more specific services. In this nonlimiting example, first tierspeech recognition servers speech recognition servers - More specifically, if second tier
speech recognition server 708 a is a speaker identification server, one or more of the firsttier recognition servers speech recognition server 708 a upon determination that a speaker identification is required. Similarly, if second tierspeech recognition server 708 b is configured for speaker verification, speech recognition servers 706 may be configured to send communication data to second tierspeech recognition server 708 b. Other configurations are also included. -
FIG. 8 is an exemplary network configuration illustrating a three-tier speech recognition configuration, similar to the network configuration fromFIG. 7 . As illustrated in the nonlimiting example ofFIG. 8 ,communications device 104 is connected to network 100. Also connected to network 100 is a first tierspeech recognition server 806. As illustrated in the previous nonlimiting examples, first tierspeech recognition server 806 may be configured to provide expedited speech recognition services via a phonetics speech to text configuration and/or an LVCSR functionality with reduced accuracy such that processing of the audio data may be completed in an expedited manner. - Also included in the nonlimiting example of
FIG. 8 is a second tierspeech recognition server 808. Second tierspeech recognition server 808 may be configured to provide speech recognition services at higher accuracy than first tierspeech recognition server 806. Similarly,FIG. 8 also includes a thirdtier recognition server 810, which may be configured to further speech recognition analytics associated with the communication. - In operation, first tier
speech recognition server 806 may be configured to receive raw data associated with a communication. First tierspeech recognition server 806 may then perform expedited speech recognition services on the received data. Second tierspeech recognition server 808 may include more thorough speech recognition functionality which may be slower in operation than first tierspeech recognition server 806, howeversecond tier server 808 may provide greater accuracy related to received data. Additionally, second tierspeech recognition server 808 may make a determination whether a third tierspeech recognition server 810 may be utilized. - Third tier
speech recognition server 810 may be configured to provide services different than that of second tierspeech recognition server 808. As a nonlimiting example, second tierspeech recognition server 808 may be configured to determined speaker confidence associated with received audio data, while a third tier speech recognition server may be configured to determine speaker emotion associated with the received audio. As such, if information regarding both speaker emotion and speaker confidence is desired, utilization of second tierspeech recognition server 808 and third tier speech recognition server 810 (as well as first tier speech recognition server 806) may be utilized. -
FIG. 9A is a flowchart illustrating an exemplary embodiment of a process that may be utilized to perform speech recognition functionality in a multi-tier speech recognition environment, such as the environment fromFIG. 6 . As illustrated in the nonlimiting example ofFIG. 9A ,call center 106 may receive raw audio data associated with a communication on a first tier speech recognition server (block 932). Callserver 106 may then determine recognition criteria (block 934). More specifically,call center 106 may be configured to determine the type of speech recognition analysis that may be performed on the received communication data. As a nonlimiting example, the receivedcommunication call center 106 may determine that the first tier speech recognition server determines speaker identification associated with communication. This indication may be received from anagent 228,analyst 230, and/orsupervisor 232, depending upon the embodiment. -
Call center 106 may then convert the received audio into a textual transcript (e.g., a phonetic transcript and/or a spoken language transcript and/or other type of transcript), as illustrated inblock 936.Call center 106 may then determine whether the audio potentially include the recognition criteria (block 938). If the received audio data does not include the recognition criteria, the process may end. If, however, first tier speech recognition server determines that the audio potentially include the recognition criteria, first tier speech recognition server can send at least a portion of the audio (which may be converted to a phonetic and/or other transcript) to second tier speech recognition server (block 938). The flowchart then proceeds to jumpblock 940, which is continued inFIG. 9B . -
FIG. 9B is a continuation from the flowchart fromFIG. 9A . As illustrated in the nonlimiting example ofFIG. 9B , a flowchart begins from ajump block 942, which is continued fromFIG. 9A . Thecall center 106 may then receive data from the first tier speech recognition server at a second tier speech recognition server (block 944). The second tier speech recognition server may then perform full recognition analytics according to the recognition criteria (block 946). - As discussed above, second tier speech recognition server may provide a more detailed speech recognition analysis of the audio data received. Similarly, some embodiments may be configured to provide a specific speech recognition analysis task such as speaker identification, speaker verification, speaker emotion, speaker confidence, and/or other types of analysis.
-
FIG. 10A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing speech recognition in a first tier upon a determination that utilization of a second speech recognition tier is desired, similar to the flowchart fromFIGS. 9A and 9B . As illustrated in the nonlimiting example ofFIG. 10A ,call center 106 may receive raw audio data at a first tier speech recognition server (block 1032). The first tier recognition server may then determine recognition criteria associated with the received audio data (block 1034). First tier speech recognition server can then begin speech to text conversion (block 1036). If at some point during the conversion, the first tier speech recognition server determines that the received valued data potentially include recognition criteria (block 1038), first tier speech recognition server can stop the conversion and send the audio data to the second tier speech recognition server (block 1040). The flowchart can then proceed to jumpblock 1042. -
FIG. 10B is a continuation of the flowchart fromFIG. 10A . As illustrated in the nonlimiting example ofFIG. 10B , fromjump block 1044, the second tier speech recognition server can receive data from the first tier speech recognition server (block 1046). Second tier speech recognition server can then perform full recognition according to the recognition criteria (block 1048). - As illustrated in this nonlimiting example, upon determination that the received audio contains one or more attributes associated with the determined speech criteria, according to the determined recognition criteria, first tier speech recognition server can send at least a portion of the data to the second tier speech recognition server. As such, full analysis of the received audio may be expedited.
-
FIG. 11A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for providing a user option for providing an agent an option to send audio data to a second tier speech recognition system, similar to the flowchart fromFIGS. 10A and 10B . As illustrated in the nonlimiting example ofFIG. 11A ,call center 106 can receive raw audio data at a first tier speech recognition server (block 1132). The first tier speech recognition server can then determine the recognition criteria associated with the received audio data (block 1134). The first tier speech recognition server can then convert the received audio into a textual format, as described above (block 1136). The first tier speech recognition server can then send at a least a portion of the conversion results to an analyst 230 (block 1138). Theanalyst 230 can then indicate whether to send the audio data to a second tier speech recognition server (block 1140). If theanalyst 230 determines that the data need not be sent to a second tier speech recognition server, the process may end. If, however, theanalyst 230 determines that further speech recognition analysis may be determined on the received audio data, the first tier speech recognition server can send at least a portion of the received audio data to a second tier speech recognition server (block 1142). Flowchart can then proceed to jumpblock 1143. -
FIG. 11B is a continuation of the flowchart fromFIG. 11A . As illustrated in the nonlimiting example ofFIG. 11B , fromjump block 1144, the second tier speech recognition server can receive at least a portion of the audio data from the first speech recognition server (block 1146). The second tier speech recognition server can then perform full recognition of the received audio data according to the recognition criteria (block 1148). As discussed above, while full recognition analytics of the received data may be performed, second tier speech recognition server may also be configured to perform specific speech recognition analytics, such as speaker identification and/or other specific speech analytics. -
FIG. 12A is a flowchart illustrating the exemplary embodiment of a process that may be utilized for determining from a plurality of second tier speech recognition servers, which server tier direct audio data, similar to the flowchart fromFIGS. 11A and 11B . As illustrated in the nonlimiting example ofFIG. 12A ,call center 106 can receive raw audio data at a first tier speech recognition server (block 1232). Thecall center 106 can then determine recognition criteria associated with the received audio data (block 1234). The first tier speech recognition server can then begin conversion of the received audio data into a textual format using phonetic speech to text conversion and/or other expedited speech to text conversion techniques (block 1236).Call center 106 may then determine whether the audio potentially include the recognition criteria (block 1238). If a determination is made that the received audio does not include the recognition criteria, the flowchart can end. If, however, a determination is made that the audio potentially include the recognition criteria, the flowchart can proceed to jumpblock 1240. -
FIG. 12B is a continuation of the flowchart fromFIG. 12A . As illustrated in the nonlimiting example ofFIG. 12B , fromjump block 1242, from the received recognition criteria,call center 106 can determine which second tier speech recognition server to be utilized for the received audio data (block 1244).Call center 106 can then receive data from the first tier speech recognition server at the determined second tier speech recognition server (block 1246). The second tier speech recognition server can then perform full recognition of the audio data according to the recognition criteria (block 1248). - It should be noted that speech analytics (i.e., the analysis of recorded speech or real-time speech) can be used to perform a variety of functions, such as automated call evaluation, call scoring, quality monitoring, quality assessment and compliance/adherence. By way of example, speech analytics can be used to compare a recorded interaction to a script (e.g., a script that the agent was to use during the interaction). In other words, speech analytics can be used to measure how well agents adhere to scripts, identify which agents are “good” sales people and which ones need additional training. As such, speech analytics can be used to find agents who do not adhere to scripts. Yet in another example, speech analytics can measure script effectiveness, identify which scripts are effective and which are not, and find, for example, the section of a script that displeases or upsets customers (e.g., based on emotion detection). As another example, compliance with various policies can be determined. Such may be in the case of, for example, the collections industry where it is a highly regulated business and agents must abide by many rules. The speech analytics of the present disclosure may identify when agents are not adhering to their scripts and guidelines. This can improve collection effectiveness and reduce corporate liability and risk.
- In this regard, various types of recording components can be used to facilitate speech analytics. Specifically, such recording components can perform one or more various functions such as receiving, capturing, intercepting and tapping of data. This can involve the use of active and/or passive recording techniques, as well as the recording of voice and/or screen data.
- It should also be noted that speech analytics can be used in conjunction with such screen data (e.g., screen data captured from an agent's workstation/PC) for evaluation, scoring, analysis, adherence and compliance purposes, for example. Such integrated functionalities improve the effectiveness and efficiency of, for example, quality assurance programs. For example, the integrated function can help companies to locate appropriate calls (and related screen interactions) for quality monitoring and evaluation. This type of “precision” monitoring improves the effectiveness and productivity of quality assurance programs.
- Another aspect that can be accomplished involves fraud detection. In this regard, various manners can be used to determine the identity of a particular speaker. In some embodiments, speech analytics can be used independently and/or in combination with other techniques for performing fraud detection. Specifically, some embodiments can involve identification of a speaker (e.g., a customer) and correlating this identification with other information to determine whether a fraudulent claim for example is being made. If such potential fraud is identified, some embodiments can provide an alert. For example, the speech analytics of the present disclosure may identify the emotions of callers. The identified emotions can be used in conjunction with identifying specific concepts to help companies spot either agents or callers/customers who are involved in fraudulent activities. Referring back to the collections example outlined above, by using emotion and concept detection, companies can identify which customers are attempting to mislead collectors into believing that they are going to pay. The earlier the company is aware of a problem account, the more recourse options they will have. Thus, the speech analytics of the present disclosure can function as an early warning system to reduce losses.
- Additionally, included in this disclosure are embodiments of integrated workforce optimization platforms, as discussed in U.S. application Ser. No. 11/359,356, filed on Feb. 22, 2006, entitled “Systems and Methods for Workforce Optimization,” Attorney Docket No 762301-1110, which is hereby incorporated by reference in its entirety. At least one embodiment of an integrated workforce optimization platform integrates: (1) Quality Monitoring/Call Recording—voice of the customer; the complete customer experience across multimedia touch points; (2) Workforce Management—strategic forecasting and scheduling that drives efficiency and adherence, aids in planning, and helps facilitate optimum staffing and service levels; (3) Performance Management—key performance indicators (KPIs) and scorecards that analyze and help identify synergies, opportunities and improvement areas; (4) e-Learning—training, new information and protocol disseminated to staff, leveraging best practice customer interactions and delivering learning to support development; and/or (5) Analytics—deliver insights from customer interactions to drive business performance. By way of example, the integrated workforce optimization process and system can include planning and establishing goals—from both an enterprise and center perspective—to ensure alignment and objectives that complement and support one another. Such planning may be complemented with forecasting and scheduling of the workforce to ensure optimum service levels. Recording and measuring performance may also be utilized, leveraging quality monitoring/call recording to assess service quality and the customer experience.
- The embodiments disclosed herein can be implemented in hardware, software, firmware, or a combination thereof. At least one embodiment, disclosed herein is implemented in software and/or firmware that is stored in a memory and that is executed by a suitable instruction execution system. If implemented in hardware, as in an alternative embodiment disclosed herein can be implemented with any or a combination of the following technologies: a discrete logic circuit) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), etc.
- One should note that the flowcharts included herein show the architecture, functionality, and operation of a possible implementation of software. In this regard, each block can be interpreted to represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order and/or not at all. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- One should note that any of the programs listed herein, which can include an ordered listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “computer-readable medium” can be any means that can contain, store, communicate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device. More specific examples (a nonexhaustive list) of the computer-readable medium could include an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM or Flash memory) (electronic), an optical fiber (optical), and a portable compact disc read-only memory (CDROM) (optical). In addition, the scope of the certain embodiments of this disclosure can include embodying the functionality described in logic embodied in hardware or software-configured mediums.
- One should also note that conditional language, such as, among others, “can,” “could,” “might,” or “may,” unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more particular embodiments or that one or more particular embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.
- It should be emphasized that the above-described embodiments are merely possible examples of implementations, merely set forth for a clear understanding of the principles of this disclosure. Many variations and modifications may be made to the above-described embodiments without departing substantially from the spirit and principles of the disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure.
Claims (24)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/924,201 US20080082334A1 (en) | 2006-09-29 | 2007-10-25 | Multi-pass speech analytics |
US13/271,681 US9171547B2 (en) | 2006-09-29 | 2011-10-12 | Multi-pass speech analytics |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/540,322 US7752043B2 (en) | 2006-09-29 | 2006-09-29 | Multi-pass speech analytics |
US11/924,201 US20080082334A1 (en) | 2006-09-29 | 2007-10-25 | Multi-pass speech analytics |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/540,322 Continuation US7752043B2 (en) | 2006-09-29 | 2006-09-29 | Multi-pass speech analytics |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/271,681 Continuation US9171547B2 (en) | 2006-09-29 | 2011-10-12 | Multi-pass speech analytics |
Publications (1)
Publication Number | Publication Date |
---|---|
US20080082334A1 true US20080082334A1 (en) | 2008-04-03 |
Family
ID=37728104
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/540,322 Active 2029-05-06 US7752043B2 (en) | 2006-09-29 | 2006-09-29 | Multi-pass speech analytics |
US11/924,201 Abandoned US20080082334A1 (en) | 2006-09-29 | 2007-10-25 | Multi-pass speech analytics |
US13/271,681 Active US9171547B2 (en) | 2006-09-29 | 2011-10-12 | Multi-pass speech analytics |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/540,322 Active 2029-05-06 US7752043B2 (en) | 2006-09-29 | 2006-09-29 | Multi-pass speech analytics |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/271,681 Active US9171547B2 (en) | 2006-09-29 | 2011-10-12 | Multi-pass speech analytics |
Country Status (2)
Country | Link |
---|---|
US (3) | US7752043B2 (en) |
CA (1) | CA2567230C (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090043581A1 (en) * | 2007-08-07 | 2009-02-12 | Aurix Limited | Methods and apparatus relating to searching of spoken audio data |
GB2451938A (en) * | 2007-08-07 | 2009-02-18 | Aurix Ltd | Methods and apparatus for searching of spoken audio data |
US20090228270A1 (en) * | 2008-03-05 | 2009-09-10 | Microsoft Corporation | Recognizing multiple semantic items from single utterance |
WO2009143135A1 (en) * | 2008-05-20 | 2009-11-26 | Calabrio, Inc. | Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms |
US20100161335A1 (en) * | 2008-12-22 | 2010-06-24 | Nortel Networks Limited | Method and system for detecting a relevant utterance |
US20100235167A1 (en) * | 2009-03-13 | 2010-09-16 | Francois Bourdon | Speech recognition learning system and method |
US20100312557A1 (en) * | 2009-06-08 | 2010-12-09 | Microsoft Corporation | Progressive application of knowledge sources in multistage speech recognition |
US20120215538A1 (en) * | 2011-02-17 | 2012-08-23 | Cisco Technology, Inc. | Performance measurement for customer contact centers |
US20140330562A1 (en) * | 2011-01-26 | 2014-11-06 | TrackThings LLC | Method and Apparatus for Obtaining Information from the Web |
Families Citing this family (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8094790B2 (en) | 2005-05-18 | 2012-01-10 | Mattersight Corporation | Method and software for training a customer service representative by analysis of a telephonic interaction between a customer and a contact center |
US8094803B2 (en) | 2005-05-18 | 2012-01-10 | Mattersight Corporation | Method and system for analyzing separated voice data of a telephonic communication between a customer and a contact center by applying a psychological behavioral model thereto |
US8433915B2 (en) * | 2006-06-28 | 2013-04-30 | Intellisist, Inc. | Selective security masking within recorded speech |
US7752043B2 (en) | 2006-09-29 | 2010-07-06 | Verint Americas Inc. | Multi-pass speech analytics |
WO2008092473A1 (en) * | 2007-01-31 | 2008-08-07 | Telecom Italia S.P.A. | Customizable method and system for emotional recognition |
WO2008106655A1 (en) * | 2007-03-01 | 2008-09-04 | Apapx, Inc. | System and method for dynamic learning |
US8886540B2 (en) * | 2007-03-07 | 2014-11-11 | Vlingo Corporation | Using speech recognition results based on an unstructured language model in a mobile communication facility application |
US10056077B2 (en) | 2007-03-07 | 2018-08-21 | Nuance Communications, Inc. | Using speech recognition results based on an unstructured language model with a music system |
US8949266B2 (en) | 2007-03-07 | 2015-02-03 | Vlingo Corporation | Multiple web-based content category searching in mobile search application |
US8880405B2 (en) * | 2007-03-07 | 2014-11-04 | Vlingo Corporation | Application text entry in a mobile environment using a speech processing facility |
US8886545B2 (en) | 2007-03-07 | 2014-11-11 | Vlingo Corporation | Dealing with switch latency in speech recognition |
US8949130B2 (en) | 2007-03-07 | 2015-02-03 | Vlingo Corporation | Internal and external speech recognition use with a mobile communication facility |
US8838457B2 (en) * | 2007-03-07 | 2014-09-16 | Vlingo Corporation | Using results of unstructured language model based speech recognition to control a system-level function of a mobile communications facility |
US8023639B2 (en) | 2007-03-30 | 2011-09-20 | Mattersight Corporation | Method and system determining the complexity of a telephonic communication received by a contact center |
US8718262B2 (en) | 2007-03-30 | 2014-05-06 | Mattersight Corporation | Method and system for automatically routing a telephonic communication base on analytic attributes associated with prior telephonic communication |
EP2196011B1 (en) | 2007-09-20 | 2017-11-08 | Unify GmbH & Co. KG | Method and communications arrangement for operating a communications connection |
US10419611B2 (en) | 2007-09-28 | 2019-09-17 | Mattersight Corporation | System and methods for determining trends in electronic communications |
US8719016B1 (en) | 2009-04-07 | 2014-05-06 | Verint Americas Inc. | Speech analytics system and system and method for determining structured speech |
US20100312469A1 (en) * | 2009-06-05 | 2010-12-09 | Telenav, Inc. | Navigation system with speech processing mechanism and method of operation thereof |
US8160877B1 (en) * | 2009-08-06 | 2012-04-17 | Narus, Inc. | Hierarchical real-time speaker recognition for biometric VoIP verification and targeting |
US8417524B2 (en) * | 2010-02-11 | 2013-04-09 | International Business Machines Corporation | Analysis of the temporal evolution of emotions in an audio interaction in a service delivery environment |
US9009040B2 (en) * | 2010-05-05 | 2015-04-14 | Cisco Technology, Inc. | Training a transcription system |
US20110307258A1 (en) * | 2010-06-10 | 2011-12-15 | Nice Systems Ltd. | Real-time application of interaction anlytics |
US9015046B2 (en) * | 2010-06-10 | 2015-04-21 | Nice-Systems Ltd. | Methods and apparatus for real-time interaction analysis in call centers |
JP5810946B2 (en) * | 2012-01-31 | 2015-11-11 | 富士通株式会社 | Specific call detection device, specific call detection method, and computer program for specific call detection |
US8781880B2 (en) | 2012-06-05 | 2014-07-15 | Rank Miner, Inc. | System, method and apparatus for voice analytics of recorded audio |
US9311914B2 (en) * | 2012-09-03 | 2016-04-12 | Nice-Systems Ltd | Method and apparatus for enhanced phonetic indexing and search |
US11568420B2 (en) * | 2012-11-21 | 2023-01-31 | Verint Americas Inc. | Analysis of customer feedback surveys |
US8537983B1 (en) | 2013-03-08 | 2013-09-17 | Noble Systems Corporation | Multi-component viewing tool for contact center agents |
US20140253666A1 (en) * | 2013-03-11 | 2014-09-11 | Qualcomm Incorporated | Intent engine for enhanced responsiveness in interactive remote communications |
US9819798B2 (en) | 2013-03-14 | 2017-11-14 | Intellisist, Inc. | Computer-implemented system and method for efficiently facilitating appointments within a call center via an automatic call distributor |
US9307084B1 (en) | 2013-04-11 | 2016-04-05 | Noble Systems Corporation | Protecting sensitive information provided by a party to a contact center |
US9225833B1 (en) | 2013-07-24 | 2015-12-29 | Noble Systems Corporation | Management system for using speech analytics to enhance contact center agent conformance |
US9407758B1 (en) | 2013-04-11 | 2016-08-02 | Noble Systems Corporation | Using a speech analytics system to control a secure audio bridge during a payment transaction |
US9456083B1 (en) | 2013-11-06 | 2016-09-27 | Noble Systems Corporation | Configuring contact center components for real time speech analytics |
US9779760B1 (en) | 2013-11-15 | 2017-10-03 | Noble Systems Corporation | Architecture for processing real time event notifications from a speech analytics system |
US9210262B1 (en) | 2013-07-24 | 2015-12-08 | Noble Systems Corporation | Using a speech analytics system to control pre-recorded scripts for debt collection calls |
US8693644B1 (en) | 2013-07-24 | 2014-04-08 | Noble Sytems Corporation | Management system for using speech analytics to enhance agent compliance for debt collection calls |
US9602665B1 (en) | 2013-07-24 | 2017-03-21 | Noble Systems Corporation | Functions and associated communication capabilities for a speech analytics component to support agent compliance in a call center |
US9674357B1 (en) | 2013-07-24 | 2017-06-06 | Noble Systems Corporation | Using a speech analytics system to control whisper audio |
US20150058006A1 (en) * | 2013-08-23 | 2015-02-26 | Xerox Corporation | Phonetic alignment for user-agent dialogue recognition |
US9191508B1 (en) | 2013-11-06 | 2015-11-17 | Noble Systems Corporation | Using a speech analytics system to offer callbacks |
US9154623B1 (en) | 2013-11-25 | 2015-10-06 | Noble Systems Corporation | Using a speech analytics system to control recording contact center calls in various contexts |
US9014364B1 (en) | 2014-03-31 | 2015-04-21 | Noble Systems Corporation | Contact center speech analytics system having multiple speech analytics engines |
US9119086B1 (en) * | 2014-05-08 | 2015-08-25 | Cellco Partnership | Evaluating 3G and voice over long term evolution voice quality |
US9654625B1 (en) | 2014-06-10 | 2017-05-16 | Tech-Friends Ltd. | Communication monitoring systems and methods |
US9953646B2 (en) | 2014-09-02 | 2018-04-24 | Belleau Technologies | Method and system for dynamic speech recognition and tracking of prewritten script |
US9659564B2 (en) * | 2014-10-24 | 2017-05-23 | Sestek Ses Ve Iletisim Bilgisayar Teknolojileri Sanayi Ticaret Anonim Sirketi | Speaker verification based on acoustic behavioral characteristics of the speaker |
US9160853B1 (en) | 2014-12-17 | 2015-10-13 | Noble Systems Corporation | Dynamic display of real time speech analytics agent alert indications in a contact center |
US9160854B1 (en) | 2014-12-17 | 2015-10-13 | Noble Systems Corporation | Reviewing call checkpoints in agent call recordings in a contact center |
US10194027B1 (en) | 2015-02-26 | 2019-01-29 | Noble Systems Corporation | Reviewing call checkpoints in agent call recording in a contact center |
US9544438B1 (en) | 2015-06-18 | 2017-01-10 | Noble Systems Corporation | Compliance management of recorded audio using speech analytics |
GB2557532A (en) * | 2015-10-21 | 2018-06-20 | Google Llc | Parameter collection and automatic dialog generation in dialog systems |
US9936066B1 (en) | 2016-03-16 | 2018-04-03 | Noble Systems Corporation | Reviewing portions of telephone call recordings in a contact center using topic meta-data records |
US9848082B1 (en) | 2016-03-28 | 2017-12-19 | Noble Systems Corporation | Agent assisting system for processing customer enquiries in a contact center |
JP6767796B2 (en) * | 2016-07-08 | 2020-10-14 | 株式会社日立情報通信エンジニアリング | Call management system and its voice recognition control method |
US10754978B2 (en) | 2016-07-29 | 2020-08-25 | Intellisist Inc. | Computer-implemented system and method for storing and retrieving sensitive information |
JP2018041230A (en) * | 2016-09-06 | 2018-03-15 | 富士通株式会社 | Reception support program, reception support method, reception support system, and information processing apparatus |
US10021245B1 (en) | 2017-05-01 | 2018-07-10 | Noble Systems Corportion | Aural communication status indications provided to an agent in a contact center |
US10832668B1 (en) * | 2017-09-19 | 2020-11-10 | Amazon Technologies, Inc. | Dynamic speech processing |
CN113257225B (en) * | 2021-05-31 | 2021-11-02 | 之江实验室 | A kind of emotional speech synthesis method and system integrating vocabulary and phoneme pronunciation features |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020046023A1 (en) * | 1995-08-18 | 2002-04-18 | Kenichi Fujii | Speech recognition system, speech recognition apparatus, and speech recognition method |
US20020046206A1 (en) * | 2000-10-17 | 2002-04-18 | Yasunari Obuchi | Method and apparatus for interpretation |
US20020156626A1 (en) * | 2001-04-20 | 2002-10-24 | Hutchison William R. | Speech recognition system |
US6487534B1 (en) * | 1999-03-26 | 2002-11-26 | U.S. Philips Corporation | Distributed client-server speech recognition system |
US20030046083A1 (en) * | 1996-11-22 | 2003-03-06 | Edward J. Devinney | User validation for information system access and transaction processing |
US20040254791A1 (en) * | 2003-03-01 | 2004-12-16 | Coifman Robert E. | Method and apparatus for improving the transcription accuracy of speech recognition software |
US20040264652A1 (en) * | 2003-06-24 | 2004-12-30 | Erhart George W. | Method and apparatus for validating agreement between textual and spoken representations of words |
US20050010422A1 (en) * | 2003-07-07 | 2005-01-13 | Canon Kabushiki Kaisha | Speech processing apparatus and method |
US20070027693A1 (en) * | 2005-07-27 | 2007-02-01 | Nec Corporation | Voice recognition system and method |
US20070198267A1 (en) * | 2002-01-04 | 2007-08-23 | Shannon Jones | Method for accessing data via voice |
US20090018834A1 (en) * | 2000-03-06 | 2009-01-15 | Cooper Robert S | Personal Virtual Assistant |
Family Cites Families (275)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3594919A (en) | 1969-09-23 | 1971-07-27 | Economy Co | Tutoring devices |
US3705271A (en) | 1971-03-26 | 1972-12-05 | Economy Co | Audio tutoring device including recording capability |
US4510351A (en) | 1982-10-28 | 1985-04-09 | At&T Bell Laboratories | ACD Management information system |
US4684349A (en) | 1984-02-15 | 1987-08-04 | Frank Ferguson | Audio-visual teaching system and method |
US4763353A (en) | 1986-02-14 | 1988-08-09 | American Telephone And Telegraph Company | Terminal based adjunct call manager for a communication system |
US4694483A (en) | 1986-06-02 | 1987-09-15 | Innings Telecom Inc. | Computerized system for routing incoming telephone calls to a plurality of agent positions |
US5008926A (en) | 1986-07-17 | 1991-04-16 | Efrat Future Technology Ltd. | Message management system |
US4924488A (en) | 1987-07-28 | 1990-05-08 | Enforcement Support Incorporated | Multiline computerized telephone monitoring system |
US4815120A (en) | 1987-07-28 | 1989-03-21 | Enforcement Support Incorporated | Computerized telephone monitoring system |
US5101402A (en) | 1988-05-24 | 1992-03-31 | Digital Equipment Corporation | Apparatus and method for realtime monitoring of network sessions in a local area network |
US4953159A (en) | 1989-01-03 | 1990-08-28 | American Telephone And Telegraph Company | Audiographics conferencing arrangement |
US5117225A (en) | 1989-05-01 | 1992-05-26 | Summit Micro Design | Computer display screen monitoring system |
US5016272A (en) | 1989-06-16 | 1991-05-14 | Stubbs James R | Home video system |
US5195086A (en) | 1990-04-12 | 1993-03-16 | At&T Bell Laboratories | Multiple call control method in a multimedia conferencing system |
US5311422A (en) | 1990-06-28 | 1994-05-10 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | General purpose architecture for intelligent computer-aided training |
US5388252A (en) | 1990-09-07 | 1995-02-07 | Eastman Kodak Company | System for transparent monitoring of processors in a network with display of screen images at a remote station for diagnosis by technical support personnel |
US5113430A (en) | 1990-10-01 | 1992-05-12 | United States Advanced Network, Inc. | Enhanced wide area audio response network |
AU9063891A (en) | 1990-11-20 | 1992-06-11 | Unifi Communications Corporation | Telephone call handling system |
US5241625A (en) | 1990-11-27 | 1993-08-31 | Farallon Computing, Inc. | Screen image sharing among heterogeneous computers |
US5239460A (en) | 1991-01-03 | 1993-08-24 | At&T Bell Laboratories | Arrangement for motivating telemarketing agents |
US5475625A (en) | 1991-01-16 | 1995-12-12 | Siemens Nixdorf Informationssysteme Aktiengesellschaft | Method and arrangement for monitoring computer manipulations |
US5381470A (en) | 1991-05-28 | 1995-01-10 | Davox Corporation | Supervisory management center with parameter testing and alerts |
US5210789A (en) | 1991-06-28 | 1993-05-11 | International Telecharge, Inc. | Interactive telephone operator terminal |
US5315711A (en) | 1991-11-01 | 1994-05-24 | Unisys Corporation | Method and apparatus for remotely and centrally controlling a plurality of host processors |
US5267865A (en) | 1992-02-11 | 1993-12-07 | John R. Lee | Interactive computer aided natural learning method and apparatus |
JPH0612288A (en) | 1992-06-29 | 1994-01-21 | Hitachi Ltd | Information processing system and its monitoring method |
GB2270581A (en) | 1992-09-15 | 1994-03-16 | Ibm | Computer workstation |
JPH0772999A (en) | 1992-10-20 | 1995-03-17 | Hewlett Packard Co <Hp> | Method and apparatus for monitoring of display screen event in screen-corresponding software application tool |
US5499291A (en) | 1993-01-14 | 1996-03-12 | At&T Corp. | Arrangement for automating call-center agent-schedule-notification and schedule-adherence functions |
DE69420096T2 (en) | 1993-09-22 | 1999-12-09 | Teknekron Infowitch Corp., Fort Worth | Telecommunication system monitoring |
US5689641A (en) | 1993-10-01 | 1997-11-18 | Vicor, Inc. | Multimedia collaboration system arrangement for routing compressed AV signal through a participant site without decompressing the AV signal |
US5347306A (en) | 1993-12-17 | 1994-09-13 | Mitsubishi Electric Research Laboratories, Inc. | Animated electronic meeting place |
US5396371A (en) | 1993-12-21 | 1995-03-07 | Dictaphone Corporation | Endless loop voice data storage and retrievable apparatus and method thereof |
US5572652A (en) | 1994-04-04 | 1996-11-05 | The United States Of America As Represented By The Secretary Of The Navy | System and method for monitoring and controlling one or more computer sites |
US5918214A (en) | 1996-10-25 | 1999-06-29 | Ipf, Inc. | System and method for finding product and service related information on the internet |
US5597312A (en) | 1994-05-04 | 1997-01-28 | U S West Technologies, Inc. | Intelligent tutoring method and system |
US5465286A (en) | 1994-05-24 | 1995-11-07 | Executone Information Systems, Inc. | Apparatus for supervising an automatic call distribution telephone system |
US5784452A (en) | 1994-06-01 | 1998-07-21 | Davox Corporation | Telephony call center with agent work groups |
US5590171A (en) | 1994-07-07 | 1996-12-31 | Bellsouth Corporation | Method and apparatus for communications monitoring |
US6130668A (en) | 1994-07-25 | 2000-10-10 | Apple Computer, Inc. | Supervisory control system for networked multimedia workstations that provides simultaneous observation of multiple remote workstations |
US5619183A (en) | 1994-09-12 | 1997-04-08 | Richard C. Ziegra | Video audio data remote system |
US5982857A (en) | 1994-10-17 | 1999-11-09 | Apropros Technology | Voice recording method and system providing context specific storage and retrieval |
JP3004883B2 (en) | 1994-10-18 | 2000-01-31 | ケイディディ株式会社 | End call detection method and apparatus and continuous speech recognition method and apparatus |
US6244758B1 (en) | 1994-11-15 | 2001-06-12 | Absolute Software Corp. | Apparatus and method for monitoring electronic devices via a global network |
US6091712A (en) | 1994-12-23 | 2000-07-18 | Applied Digital Access, Inc. | Method and apparatus for storing and retrieving performance data collected by a network interface unit |
US5742670A (en) | 1995-01-09 | 1998-04-21 | Ncr Corporation | Passive telephone monitor to control collaborative systems |
US5696906A (en) | 1995-03-09 | 1997-12-09 | Continental Cablevision, Inc. | Telecommunicaion user account management system and method |
US5918222A (en) | 1995-03-17 | 1999-06-29 | Kabushiki Kaisha Toshiba | Information disclosing apparatus and multi-modal information input/output system |
DE69636239T2 (en) | 1995-04-24 | 2007-05-10 | International Business Machines Corp. | A method and apparatus for skill-based routing in a call center |
US5721842A (en) | 1995-08-25 | 1998-02-24 | Apex Pc Solutions, Inc. | Interconnection system for viewing and controlling remotely connected computers with on-screen video overlay for controlling of the interconnection switch |
US5748499A (en) | 1995-09-19 | 1998-05-05 | Sony Corporation | Computer graphics data recording and playback system with a VCR-based graphic user interface |
US5884032A (en) | 1995-09-25 | 1999-03-16 | The New Brunswick Telephone Company, Limited | System for coordinating communications via customer contact channel changing system using call centre for setting up the call between customer and an available help agent |
US6122668A (en) | 1995-11-02 | 2000-09-19 | Starlight Networks | Synchronization of audio and video signals in a live multicast in a LAN |
US5717879A (en) | 1995-11-03 | 1998-02-10 | Xerox Corporation | System for the capture and replay of temporal data representing collaborative activities |
US5778182A (en) | 1995-11-07 | 1998-07-07 | At&T Corp. | Usage management system |
US6052454A (en) | 1996-01-16 | 2000-04-18 | Global Tel*Link Corp. | Telephone apparatus with recording of phone conversations on massive storage |
US5826014A (en) | 1996-02-06 | 1998-10-20 | Network Engineering Software | Firewall system for protecting network elements connected to a public network |
US6225993B1 (en) | 1996-04-22 | 2001-05-01 | Sun Microsystems, Inc. | Video on demand applet method and apparatus for inclusion of motion video in multimedia documents |
US5727950A (en) | 1996-05-22 | 1998-03-17 | Netsage Corporation | Agent based instruction system and method |
US6018619A (en) | 1996-05-24 | 2000-01-25 | Microsoft Corporation | Method, system and apparatus for client-side usage tracking of information server systems |
US5790798A (en) | 1996-05-31 | 1998-08-04 | Witness Systems, Inc. | Method and apparatus for simultaneously monitoring computer user screen and telephone activity from a remote location |
US20030144900A1 (en) | 2002-01-28 | 2003-07-31 | Whitmer Michael L. | Method and system for improving enterprise performance |
US5907680A (en) | 1996-06-24 | 1999-05-25 | Sun Microsystems, Inc. | Client-side, server-side and collaborative spell check of URL's |
US5862330A (en) | 1996-07-16 | 1999-01-19 | Lucent Technologies Inc. | Technique for obtaining and exchanging information on wolrd wide web |
US6157808A (en) | 1996-07-17 | 2000-12-05 | Gpu, Inc. | Computerized employee certification and training system |
US5809247A (en) | 1996-07-22 | 1998-09-15 | Intel Corporation | Method and apparatus for guided touring of internet/intranet websites |
US5757644A (en) | 1996-07-25 | 1998-05-26 | Eis International, Inc. | Voice interactive call center training method using actual screens and screen logic |
US5933811A (en) | 1996-08-20 | 1999-08-03 | Paul D. Angles | System and method for delivering customized advertisements within interactive communication systems |
US6014134A (en) | 1996-08-23 | 2000-01-11 | U S West, Inc. | Network-based intelligent tutoring system |
US5923746A (en) | 1996-09-18 | 1999-07-13 | Rockwell International Corp. | Call recording system and method for use with a telephonic switch |
NZ334584A (en) | 1996-09-25 | 2001-02-23 | Sylvan Learning Systems Inc | Automated testing and electronic instructional delivery and student management system |
GB9620082D0 (en) | 1996-09-26 | 1996-11-13 | Eyretel Ltd | Signal monitoring apparatus |
US5944791A (en) | 1996-10-04 | 1999-08-31 | Contigo Software Llc | Collaborative web browser |
US6487195B1 (en) | 1996-10-23 | 2002-11-26 | Ncr Corporation | Collaborative network navigation synchronization mechanism |
US5809250A (en) | 1996-10-23 | 1998-09-15 | Intel Corporation | Methods for creating and sharing replayable modules representive of Web browsing session |
US6039575A (en) | 1996-10-24 | 2000-03-21 | National Education Corporation | Interactive learning system with pretest |
US5948061A (en) | 1996-10-29 | 1999-09-07 | Double Click, Inc. | Method of delivery, targeting, and measuring advertising over networks |
US5990852A (en) | 1996-10-31 | 1999-11-23 | Fujitsu Limited | Display screen duplication system and method |
US5864772A (en) | 1996-12-23 | 1999-01-26 | Schlumberger Technology Corporation | Apparatus, system and method to transmit and display acquired well data in near real time at a remote location |
US5917489A (en) | 1997-01-31 | 1999-06-29 | Microsoft Corporation | System and method for creating, editing, and distributing rules for processing electronic messages |
US6560328B1 (en) | 1997-04-03 | 2003-05-06 | Genesys Telecommunications Laboratories, Inc. | Voice extensions in a call-in center employing virtual restructuring for computer telephony integrated functionality |
US5978648A (en) | 1997-03-06 | 1999-11-02 | Forte Systems, Inc. | Interactive multimedia performance assessment system and process for use by students, educators and administrators |
US5796952A (en) | 1997-03-21 | 1998-08-18 | Dot Com Development, Inc. | Method and apparatus for tracking client interaction with a network resource and creating client profiles and resource database |
US6301573B1 (en) | 1997-03-21 | 2001-10-09 | Knowlagent, Inc. | Recurrent training system |
US6078894A (en) | 1997-03-28 | 2000-06-20 | Clawson; Jeffrey J. | Method and system for evaluating the performance of emergency medical dispatchers |
US6578077B1 (en) | 1997-05-27 | 2003-06-10 | Novell, Inc. | Traffic monitoring tool for bandwidth management |
US6171109B1 (en) | 1997-06-18 | 2001-01-09 | Adin Research, Inc. | Method for generating a multi-strata model and an intellectual information processing device |
US6282548B1 (en) | 1997-06-21 | 2001-08-28 | Alexa Internet | Automatically generate and displaying metadata as supplemental information concurrently with the web page, there being no link between web page and metadata |
CA2401676C (en) | 1997-06-25 | 2008-09-16 | Samsung Electronics Co., Ltd. | Browser based command and control home network |
US6014647A (en) | 1997-07-08 | 2000-01-11 | Nizzari; Marcia M. | Customer interaction tracking |
US5958016A (en) | 1997-07-13 | 1999-09-28 | Bell Atlantic Network Services, Inc. | Internet-web link for access to intelligent network service control |
US6076099A (en) | 1997-09-09 | 2000-06-13 | Chen; Thomas C. H. | Method for configurable intelligent-agent-based wireless communication system |
US5964836A (en) | 1997-09-11 | 1999-10-12 | International Business Machines Corporation | Apparatus, methods and computer program products for managing web-page-embedded sessions with a host-based application |
US5991373A (en) | 1997-09-15 | 1999-11-23 | Teknekron Infoswitch Corporation | Reproduction of a voice and video session |
US6035332A (en) | 1997-10-06 | 2000-03-07 | Ncr Corporation | Method for monitoring user interactions with web pages from web server using data and command lists for maintaining information visited and issued by participants |
US6418471B1 (en) | 1997-10-06 | 2002-07-09 | Ncr Corporation | Method for recording and reproducing the browsing activities of an individual web browser |
US6546405B2 (en) | 1997-10-23 | 2003-04-08 | Microsoft Corporation | Annotating temporally-dimensioned multimedia content |
US6351467B1 (en) | 1997-10-27 | 2002-02-26 | Hughes Electronics Corporation | System and method for multicasting multimedia content |
US6009429A (en) | 1997-11-13 | 1999-12-28 | International Business Machines Corporation | HTML guided web tour |
US5987466A (en) | 1997-11-25 | 1999-11-16 | International Business Machines Corporation | Presenting web pages with discrete, browser-controlled complexity levels |
US6286046B1 (en) | 1997-12-22 | 2001-09-04 | International Business Machines Corporation | Method of recording and measuring e-business sessions on the world wide web |
US6005932A (en) | 1997-12-24 | 1999-12-21 | Rockwell Semiconductor Systems Inc. | Dynamic schedule profiler for ACD |
US6195679B1 (en) | 1998-01-06 | 2001-02-27 | Netscape Communications Corporation | Browsing session recording playback and editing system for generating user defined paths and allowing users to mark the priority of items in the paths |
JP3371791B2 (en) | 1998-01-29 | 2003-01-27 | ヤマハ株式会社 | Music training system and music training device, and recording medium on which music training program is recorded |
US6151622A (en) | 1998-02-02 | 2000-11-21 | International Business Machines Corp. | Method and system for portably enabling view synchronization over the world-wide web using frame hierarchies |
US6144991A (en) | 1998-02-19 | 2000-11-07 | Telcordia Technologies, Inc. | System and method for managing interactions between users in a browser-based telecommunications network |
US6230197B1 (en) | 1998-09-11 | 2001-05-08 | Genesys Telecommunications Laboratories, Inc. | Method and apparatus for rules-based storage and retrieval of multimedia interactions within a communication center |
US6138139A (en) | 1998-10-29 | 2000-10-24 | Genesys Telecommunications Laboraties, Inc. | Method and apparatus for supporting diverse interaction paths within a multimedia communication center |
US6038544A (en) | 1998-02-26 | 2000-03-14 | Teknekron Infoswitch Corporation | System and method for determining the performance of a user responding to a call |
US20030154072A1 (en) * | 1998-03-31 | 2003-08-14 | Scansoft, Inc., A Delaware Corporation | Call analysis |
US20010043697A1 (en) | 1998-05-11 | 2001-11-22 | Patrick M. Cox | Monitoring of and remote access to call center activity |
US6154771A (en) | 1998-06-01 | 2000-11-28 | Mediastra, Inc. | Real-time receipt, decompression and play of compressed streaming video/hypervideo; with thumbnail display of past scenes and with replay, hyperlinking and/or recording permissively intiated retrospectively |
US6347374B1 (en) | 1998-06-05 | 2002-02-12 | Intrusion.Com, Inc. | Event detection |
US6100891A (en) | 1998-06-09 | 2000-08-08 | Teledirect International, Inc. | Call center agent interface and development tool |
CN1139254C (en) | 1998-06-26 | 2004-02-18 | 通用仪器公司 | Terminal for composing and presenting MPEG-4 video programs |
US6286030B1 (en) | 1998-07-10 | 2001-09-04 | Sap Aktiengesellschaft | Systems and methods for recording and visually recreating sessions in a client-server environment |
US6122665A (en) | 1998-08-26 | 2000-09-19 | Sts Software System Ltd. | Communication management system for computer network-based telephones |
FR2782875B1 (en) | 1998-08-27 | 2000-11-03 | France Telecom | TELEPHONE DEVICE FOR PRISON |
US6493758B1 (en) | 1998-09-08 | 2002-12-10 | Microsoft Corporation | Offline viewing of internet content with a mobile device |
US6353851B1 (en) | 1998-12-28 | 2002-03-05 | Lucent Technologies Inc. | Method and apparatus for sharing asymmetric information and services in simultaneously viewed documents on a communication system |
US6411989B1 (en) | 1998-12-28 | 2002-06-25 | Lucent Technologies Inc. | Apparatus and method for sharing information in simultaneously viewed documents on a communication system |
US6360250B1 (en) | 1998-12-28 | 2002-03-19 | Lucent Technologies Inc. | Apparatus and method for sharing information in simultaneously viewed documents on a communication system |
US6236977B1 (en) | 1999-01-04 | 2001-05-22 | Realty One, Inc. | Computer implemented marketing system |
US6301462B1 (en) | 1999-01-15 | 2001-10-09 | Unext. Com | Online collaborative apprenticeship |
US6370547B1 (en) | 1999-04-21 | 2002-04-09 | Union Oil Company Of California | Database correlation method |
US6606657B1 (en) | 1999-06-22 | 2003-08-12 | Comverse, Ltd. | System and method for processing and presenting internet usage information |
US6288753B1 (en) | 1999-07-07 | 2001-09-11 | Corrugated Services Corp. | System and method for live interactive distance learning |
US6289340B1 (en) | 1999-08-03 | 2001-09-11 | Ixmatch, Inc. | Consultant matching system and method for selecting candidates from a candidate pool by adjusting skill values |
US6665644B1 (en) | 1999-08-10 | 2003-12-16 | International Business Machines Corporation | Conversational data mining |
US6275806B1 (en) | 1999-08-31 | 2001-08-14 | Andersen Consulting, Llp | System method and article of manufacture for detecting emotion in voice signals by utilizing statistics for voice signal parameters |
US6697457B2 (en) | 1999-08-31 | 2004-02-24 | Accenture Llp | Voice messaging system that organizes voice messages based on detected emotion |
US6772396B1 (en) | 1999-10-07 | 2004-08-03 | Microsoft Corporation | Content distribution system for network environments |
US6823384B1 (en) | 1999-10-15 | 2004-11-23 | James Wilson | Methods and apparatus for securely collecting customer service agent data in a multi-tenant environment |
US6792575B1 (en) | 1999-10-21 | 2004-09-14 | Equilibrium Technologies | Automated processing and delivery of media to web servers |
US7725307B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Query engine for processing voice based queries including semantic decoding |
US6901438B1 (en) | 1999-11-12 | 2005-05-31 | Bmc Software | System selects a best-fit form or URL in an originating web page as a target URL for replaying a predefined path through the internet |
US6535909B1 (en) | 1999-11-18 | 2003-03-18 | Contigo Software, Inc. | System and method for record and playback of collaborative Web browsing session |
US7613695B1 (en) | 1999-12-06 | 2009-11-03 | Reed Elsevier Inc. | Relationship management system that provides an indication of users having a relationship with a specified contact |
US6674447B1 (en) | 1999-12-06 | 2004-01-06 | Oridus, Inc. | Method and apparatus for automatically recording snapshots of a computer screen during a computer session for later playback |
US6959078B1 (en) | 2000-01-24 | 2005-10-25 | Verint Systems Inc. | Apparatus and method for monitoring and adapting to environmental factors within a contact center |
IL141002A0 (en) | 2000-01-24 | 2002-02-10 | Comverse Infosys Inc | Open storage portal apparatus and method to access contact center information |
US6724887B1 (en) | 2000-01-24 | 2004-04-20 | Verint Systems, Inc. | Method and system for analyzing customer communications with a contact center |
US6810414B1 (en) | 2000-02-04 | 2004-10-26 | Dennis A. Brittain | System and methods for easy-to-use periodic network data capture engine with automatic target data location, extraction and storage |
AU3689901A (en) | 2000-02-10 | 2001-08-20 | Involve Technology Llc | System for creating and maintaining a database of information utilizing user opinions |
US6542602B1 (en) | 2000-02-14 | 2003-04-01 | Nice Systems Ltd. | Telephone call monitoring system |
US6324282B1 (en) | 2000-03-02 | 2001-11-27 | Knowlagent, Inc. | Method and system for delivery of individualized training to call center agents |
US6775377B2 (en) | 2001-09-10 | 2004-08-10 | Knowlagent, Inc. | Method and system for delivery of individualized training to call center agents |
AU2001245426A1 (en) | 2000-03-03 | 2001-09-17 | Lawrence R. Jones | Picture communications system and associated network services |
US6683633B2 (en) | 2000-03-20 | 2004-01-27 | Incontext Enterprises, Inc. | Method and system for accessing information |
US6721734B1 (en) | 2000-04-18 | 2004-04-13 | Claritech Corporation | Method and apparatus for information management using fuzzy typing |
US6697858B1 (en) | 2000-08-14 | 2004-02-24 | Telephony@Work | Call center |
EP1189161A1 (en) | 2000-09-13 | 2002-03-20 | iMediation, S.A. | A method and system for managing network-based partner relationships |
US7287071B2 (en) | 2000-09-28 | 2007-10-23 | Vignette Corporation | Transaction management system |
US20020065911A1 (en) | 2000-10-03 | 2002-05-30 | Von Klopp Ana H. | HTTP transaction monitor with edit and replay capacity |
US7451085B2 (en) * | 2000-10-13 | 2008-11-11 | At&T Intellectual Property Ii, L.P. | System and method for providing a compensated speech recognition model for speech recognition |
US6751614B1 (en) | 2000-11-09 | 2004-06-15 | Satyam Computer Services Limited Of Mayfair Centre | System and method for topic-based document analysis for information filtering |
US20020065912A1 (en) | 2000-11-30 | 2002-05-30 | Catchpole Lawrence W. | Web session collaboration |
US9058416B2 (en) | 2000-12-11 | 2015-06-16 | Peter K. Trzyna | System and method for detecting and reporting online activity using real-time content-based network monitoring |
US20020143925A1 (en) | 2000-12-29 | 2002-10-03 | Ncr Corporation | Identifying web-log data representing a single user session |
US7739115B1 (en) | 2001-02-15 | 2010-06-15 | West Corporation | Script compliance and agent feedback |
US7039166B1 (en) | 2001-03-05 | 2006-05-02 | Verizon Corporate Services Group Inc. | Apparatus and method for visually representing behavior of a user of an automated response system |
US6823054B1 (en) | 2001-03-05 | 2004-11-23 | Verizon Corporate Services Group Inc. | Apparatus and method for analyzing an automated response system |
US7506047B2 (en) | 2001-03-30 | 2009-03-17 | Bmc Software, Inc. | Synthetic transaction monitor with replay capability |
US6944660B2 (en) | 2001-05-04 | 2005-09-13 | Hewlett-Packard Development Company, L.P. | System and method for monitoring browser event activities |
EP1256937B1 (en) | 2001-05-11 | 2006-11-02 | Sony France S.A. | Emotion recognition method and device |
US6801604B2 (en) * | 2001-06-25 | 2004-10-05 | International Business Machines Corporation | Universal IP-based and scalable architectures across conversational applications using web services for speech and audio processing resources |
US20040100507A1 (en) | 2001-08-24 | 2004-05-27 | Omri Hayner | System and method for capturing browser sessions and user actions |
US6738456B2 (en) | 2001-09-07 | 2004-05-18 | Ronco Communications And Electronics, Inc. | School observation and supervisory system |
US6870916B2 (en) | 2001-09-14 | 2005-03-22 | Lucent Technologies Inc. | Targeted and intelligent multimedia conference establishment services |
JP3644955B2 (en) | 2001-09-27 | 2005-05-11 | 松下電器産業株式会社 | Conversation device, conversation master device, conversation slave device, conversation control method, and conversation control program |
US20030079020A1 (en) | 2001-10-23 | 2003-04-24 | Christophe Gourraud | Method, system and service provider for IP media program transfer-and-viewing-on-demand |
US7613717B1 (en) | 2001-10-26 | 2009-11-03 | Teradata Us, Inc. | Automated system for rating customer feedback |
US6965886B2 (en) | 2001-11-01 | 2005-11-15 | Actimize Ltd. | System and method for analyzing and utilizing data, by executing complex analytical models in real time |
US6801618B2 (en) | 2002-02-08 | 2004-10-05 | Etalk Corporation | System and method for implementing recording plans using a session manager |
US6914975B2 (en) | 2002-02-21 | 2005-07-05 | Sbc Properties, L.P. | Interactive dialog-based training method |
US8239197B2 (en) | 2002-03-28 | 2012-08-07 | Intellisist, Inc. | Efficient conversion of voice messages into text |
US7076430B1 (en) | 2002-05-16 | 2006-07-11 | At&T Corp. | System and method of providing conversational visual prosody for talking heads |
WO2004012431A1 (en) | 2002-07-29 | 2004-02-05 | British Telecommunications Public Limited Company | Improvements in or relating to information provision for call centres |
EP1556854B1 (en) | 2002-09-27 | 2011-05-04 | Callminer, Inc. | Method for statistical analysis of speech |
US8055503B2 (en) | 2002-10-18 | 2011-11-08 | Siemens Enterprise Communications, Inc. | Methods and apparatus for audio data analysis and data mining using speech recognition |
EP1565907A4 (en) | 2002-10-18 | 2006-01-18 | Ser Solutions Inc | Methods and apparatus for audio data monitoring and evaluation using speech recognition |
US20040098265A1 (en) | 2002-11-05 | 2004-05-20 | Sean Kelly | Dialog management system |
US7606714B2 (en) | 2003-02-11 | 2009-10-20 | Microsoft Corporation | Natural language classification within an automated response system |
US20040162724A1 (en) | 2003-02-11 | 2004-08-19 | Jeffrey Hill | Management of conversations |
AU2003276661A1 (en) | 2003-11-05 | 2005-05-26 | Nice Systems Ltd. | Apparatus and method for event-driven content analysis |
US20050138560A1 (en) | 2003-12-18 | 2005-06-23 | Kuo-Chun Lee | Method and apparatus for broadcasting live personal performances over the internet |
US20080235018A1 (en) | 2004-01-20 | 2008-09-25 | Koninklikke Philips Electronic,N.V. | Method and System for Determing the Topic of a Conversation and Locating and Presenting Related Content |
US9076343B2 (en) | 2004-04-06 | 2015-07-07 | International Business Machines Corporation | Self-service system for education |
US8805717B2 (en) | 2004-08-31 | 2014-08-12 | Hartford Fire Insurance Company | Method and system for improving performance of customer service representatives |
AU2006211111A1 (en) | 2005-02-07 | 2006-08-10 | Nice Systems Ltd. | Upgrading performance using aggregated information shared between management systems |
US7720214B2 (en) | 2005-02-22 | 2010-05-18 | International Business Machines Corporation | Call center study method and system |
US8032823B2 (en) | 2005-04-15 | 2011-10-04 | Carnegie Mellon University | Intent-based information processing and updates |
US8568144B2 (en) | 2005-05-09 | 2013-10-29 | Altis Avante Corp. | Comprehension instruction system and method |
US7995717B2 (en) | 2005-05-18 | 2011-08-09 | Mattersight Corporation | Method and system for analyzing separated voice data of a telephonic communication between a customer and a contact center by applying a psychological behavioral model thereto |
US7940897B2 (en) | 2005-06-24 | 2011-05-10 | American Express Travel Related Services Company, Inc. | Word recognition system and method for customer and employee assessment |
US9300790B2 (en) | 2005-06-24 | 2016-03-29 | Securus Technologies, Inc. | Multi-party conversation analyzer and logger |
US20070016580A1 (en) | 2005-07-15 | 2007-01-18 | International Business Machines Corporation | Extracting information about references to entities rom a plurality of electronic documents |
US20070043608A1 (en) | 2005-08-22 | 2007-02-22 | Recordant, Inc. | Recorded customer interactions and training system, method and computer program product |
NO20054720L (en) | 2005-10-13 | 2007-04-16 | Fast Search & Transfer Asa | Information access with user-driven metadata feedback |
US20070198330A1 (en) | 2006-02-22 | 2007-08-23 | Shmuel Korenblit | Integrated contact center systems for facilitating contact center coaching |
US8396741B2 (en) | 2006-02-22 | 2013-03-12 | 24/7 Customer, Inc. | Mining interactions to manage customer experience throughout a customer service lifecycle |
US8108237B2 (en) | 2006-02-22 | 2012-01-31 | Verint Americas, Inc. | Systems for integrating contact center monitoring, training and scheduling |
JP4870448B2 (en) | 2006-02-23 | 2012-02-08 | 株式会社日立製作所 | Information processing apparatus, customer needs analysis method, and program |
US7593522B2 (en) | 2006-03-09 | 2009-09-22 | At&T Intellectual Property I, L.P. | Call center user interface and methods of using same |
US20070287477A1 (en) * | 2006-06-12 | 2007-12-13 | Available For Licensing | Mobile device with shakeable snow rendering |
US7865510B2 (en) | 2006-07-12 | 2011-01-04 | LitCentral, Inc | Internet user-accessible database |
WO2008014182A2 (en) | 2006-07-24 | 2008-01-31 | Chacha Search, Inc. | Method, system, and computer readable storage for podcasting and video training in an information search system |
US7752043B2 (en) | 2006-09-29 | 2010-07-06 | Verint Americas Inc. | Multi-pass speech analytics |
US7570755B2 (en) | 2006-09-29 | 2009-08-04 | Verint Americas Inc. | Routine communication sessions for recording |
US7991613B2 (en) | 2006-09-29 | 2011-08-02 | Verint Americas Inc. | Analyzing audio components and generating text with integrated additional session information |
US8050923B2 (en) | 2006-09-29 | 2011-11-01 | Verint Americas, Inc. | Automated utterance search |
US20080091423A1 (en) | 2006-10-13 | 2008-04-17 | Shourya Roy | Generation of domain models from noisy transcriptions |
US7577246B2 (en) | 2006-12-20 | 2009-08-18 | Nice Systems Ltd. | Method and system for automatic quality evaluation |
US8571853B2 (en) | 2007-02-11 | 2013-10-29 | Nice Systems Ltd. | Method and system for laughter detection |
US7689624B2 (en) | 2007-03-01 | 2010-03-30 | Microsoft Corporation | Graph-based search leveraging sentiment analysis of user comments |
US20080249764A1 (en) | 2007-03-01 | 2008-10-09 | Microsoft Corporation | Smart Sentiment Classifier for Product Reviews |
US7996210B2 (en) | 2007-04-24 | 2011-08-09 | The Research Foundation Of The State University Of New York | Large-scale sentiment analysis |
US8200527B1 (en) | 2007-04-25 | 2012-06-12 | Convergys Cmg Utah, Inc. | Method for prioritizing and presenting recommendations regarding organizaion's customer care capabilities |
US20080281581A1 (en) | 2007-05-07 | 2008-11-13 | Sparta, Inc. | Method of identifying documents with similar properties utilizing principal component analysis |
CA2640779A1 (en) | 2007-10-02 | 2009-04-02 | Neurolanguage Corporation | Computer-based language training work plan creation with specialized english materials |
US8209209B2 (en) | 2007-10-02 | 2012-06-26 | Incontact, Inc. | Providing work, training, and incentives to company representatives in contact handling systems |
US8166013B2 (en) | 2007-11-05 | 2012-04-24 | Intuit Inc. | Method and system for crawling, mapping and extracting information associated with a business using heuristic and semantic analysis |
US8046220B2 (en) | 2007-11-28 | 2011-10-25 | Nuance Communications, Inc. | Systems and methods to index and search voice sites |
US8417713B1 (en) | 2007-12-05 | 2013-04-09 | Google Inc. | Sentiment detection as a ranking signal for reviewable entities |
US8010539B2 (en) | 2008-01-25 | 2011-08-30 | Google Inc. | Phrase based snippet generation |
US9122749B2 (en) | 2009-02-04 | 2015-09-01 | Popular Metrics, Inc. | Internet based system and method for wagering on an artist |
US9323836B2 (en) | 2008-02-11 | 2016-04-26 | Popular Metrics, Inc. | Internet based method and system for ranking artists using a popularity profile |
US7925743B2 (en) | 2008-02-29 | 2011-04-12 | Networked Insights, Llc | Method and system for qualifying user engagement with a website |
US20090228428A1 (en) | 2008-03-07 | 2009-09-10 | International Business Machines Corporation | Solution for augmenting a master data model with relevant data elements extracted from unstructured data sources |
GB2458461A (en) | 2008-03-17 | 2009-09-23 | Kai Yu | Spoken language learning system |
US8463594B2 (en) | 2008-03-21 | 2013-06-11 | Sauriel Llc | System and method for analyzing text using emotional intelligence factors |
US8117207B2 (en) | 2008-04-18 | 2012-02-14 | Biz360 Inc. | System and methods for evaluating feature opinions for products, services, and entities |
US7913063B1 (en) | 2008-05-02 | 2011-03-22 | Verint Americas Inc. | System and method for performance based call distribution |
US8543393B2 (en) | 2008-05-20 | 2013-09-24 | Calabrio, Inc. | Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms |
US8140330B2 (en) | 2008-06-13 | 2012-03-20 | Robert Bosch Gmbh | System and method for detecting repeated patterns in dialog systems |
US8219555B1 (en) | 2008-06-13 | 2012-07-10 | Ustringer LLC | Method and apparatus for distributing content |
JP5226401B2 (en) | 2008-06-25 | 2013-07-03 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Apparatus and method for supporting retrieval of document data |
US8965765B2 (en) | 2008-09-19 | 2015-02-24 | Microsoft Corporation | Structured models of repetition for speech recognition |
US8886663B2 (en) | 2008-09-20 | 2014-11-11 | Securus Technologies, Inc. | Multi-party conversation analyzer and logger |
US8520808B2 (en) | 2008-10-08 | 2013-08-27 | Synchronoss Technologies | System and method for robust evaluation of the user experience in automated spoken dialog systems |
TW201042987A (en) | 2008-10-17 | 2010-12-01 | Commw Intellectual Property Holdings Inc | Intuitive voice navigation |
US8750489B2 (en) | 2008-10-23 | 2014-06-10 | International Business Machines Corporation | System and method for automatic call segmentation at call center |
US8644488B2 (en) | 2008-10-27 | 2014-02-04 | Nuance Communications, Inc. | System and method for automatically generating adaptive interaction logs from customer interaction text |
US20100119053A1 (en) | 2008-11-13 | 2010-05-13 | Buzzient, Inc. | Analytic measurement of online social media content |
US8606815B2 (en) | 2008-12-09 | 2013-12-10 | International Business Machines Corporation | Systems and methods for analyzing electronic text |
US8756065B2 (en) | 2008-12-24 | 2014-06-17 | At&T Intellectual Property I, L.P. | Correlated call analysis for identified patterns in call transcriptions |
JP2010181993A (en) | 2009-02-04 | 2010-08-19 | Kddi Corp | Evaluation analysis server, method, and program for evaluating text file containing pictorial symbol |
US8798255B2 (en) | 2009-03-31 | 2014-08-05 | Nice Systems Ltd | Methods and apparatus for deep interaction analysis |
JP5495602B2 (en) | 2009-04-02 | 2014-05-21 | オリンパスイメージング株式会社 | Imaging apparatus and imaging method |
US8166032B2 (en) | 2009-04-09 | 2012-04-24 | MarketChorus, Inc. | System and method for sentiment-based text classification and relevancy ranking |
US8370155B2 (en) | 2009-04-23 | 2013-02-05 | International Business Machines Corporation | System and method for real time support for agents in contact center environments |
US8054964B2 (en) | 2009-04-30 | 2011-11-08 | Avaya Inc. | System and method for detecting emotions at different steps in a communication |
US8494133B2 (en) | 2009-06-24 | 2013-07-23 | Nexidia Inc. | Enterprise speech intelligence analysis |
US20100332287A1 (en) | 2009-06-24 | 2010-12-30 | International Business Machines Corporation | System and method for real-time prediction of customer satisfaction |
US8463606B2 (en) | 2009-07-13 | 2013-06-11 | Genesys Telecommunications Laboratories, Inc. | System for analyzing interactions and reporting analytic results to human-operated and system interfaces in real time |
US8533208B2 (en) | 2009-09-28 | 2013-09-10 | Ebay Inc. | System and method for topic extraction and opinion mining |
US8880537B2 (en) | 2009-10-19 | 2014-11-04 | Gil Fuchs | System and method for use of semantic understanding in storage, searching and providing of data or other content information |
US8626753B1 (en) | 2009-11-19 | 2014-01-07 | Anshu Aggarwal | Personalization search engine |
US20130297581A1 (en) | 2009-12-01 | 2013-11-07 | Topsy Labs, Inc. | Systems and methods for customized filtering and analysis of social media content collected over social networks |
US8417524B2 (en) | 2010-02-11 | 2013-04-09 | International Business Machines Corporation | Analysis of the temporal evolution of emotions in an audio interaction in a service delivery environment |
US8412530B2 (en) | 2010-02-21 | 2013-04-02 | Nice Systems Ltd. | Method and apparatus for detection of sentiment in automated transcriptions |
US20110216905A1 (en) | 2010-03-05 | 2011-09-08 | Nexidia Inc. | Channel compression |
US8620849B2 (en) | 2010-03-10 | 2013-12-31 | Lockheed Martin Corporation | Systems and methods for facilitating open source intelligence gathering |
US20110238670A1 (en) | 2010-03-23 | 2011-09-29 | Microsoft Corporation | Crowd-sourcing and contextual reclassification of rated content |
US8880559B2 (en) | 2010-04-02 | 2014-11-04 | Brian Bartell | Location activity search engine computer system |
US8306814B2 (en) | 2010-05-11 | 2012-11-06 | Nice-Systems Ltd. | Method for speaker source classification |
US9015046B2 (en) | 2010-06-10 | 2015-04-21 | Nice-Systems Ltd. | Methods and apparatus for real-time interaction analysis in call centers |
US20120130771A1 (en) | 2010-11-18 | 2012-05-24 | Kannan Pallipuram V | Chat Categorization and Agent Performance Modeling |
US8731918B2 (en) | 2011-02-23 | 2014-05-20 | Nicesystems Ltd. | Method and apparatus for automatic correlation of multi-channel interactions |
US8531501B2 (en) | 2011-03-01 | 2013-09-10 | Nice-Systems Ltd. | System and method for assisting an agent in a contact center |
WO2012134877A2 (en) | 2011-03-25 | 2012-10-04 | Educational Testing Service | Computer-implemented systems and methods evaluating prosodic features of speech |
US20120253792A1 (en) | 2011-03-30 | 2012-10-04 | Nec Laboratories America, Inc. | Sentiment Classification Based on Supervised Latent N-Gram Analysis |
US20130018875A1 (en) | 2011-07-11 | 2013-01-17 | Lexxe Pty Ltd | System and method for ordering semantic sub-keys utilizing superlative adjectives |
US8463595B1 (en) | 2012-03-06 | 2013-06-11 | Reputation.Com, Inc. | Detailed sentiment analysis |
US10304036B2 (en) | 2012-05-07 | 2019-05-28 | Nasdaq, Inc. | Social media profiling for one or more authors using one or more social media platforms |
-
2006
- 2006-09-29 US US11/540,322 patent/US7752043B2/en active Active
- 2006-11-07 CA CA2567230A patent/CA2567230C/en not_active Expired - Fee Related
-
2007
- 2007-10-25 US US11/924,201 patent/US20080082334A1/en not_active Abandoned
-
2011
- 2011-10-12 US US13/271,681 patent/US9171547B2/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020046023A1 (en) * | 1995-08-18 | 2002-04-18 | Kenichi Fujii | Speech recognition system, speech recognition apparatus, and speech recognition method |
US20030046083A1 (en) * | 1996-11-22 | 2003-03-06 | Edward J. Devinney | User validation for information system access and transaction processing |
US6487534B1 (en) * | 1999-03-26 | 2002-11-26 | U.S. Philips Corporation | Distributed client-server speech recognition system |
US20090018834A1 (en) * | 2000-03-06 | 2009-01-15 | Cooper Robert S | Personal Virtual Assistant |
US20020046206A1 (en) * | 2000-10-17 | 2002-04-18 | Yasunari Obuchi | Method and apparatus for interpretation |
US20020156626A1 (en) * | 2001-04-20 | 2002-10-24 | Hutchison William R. | Speech recognition system |
US20070198267A1 (en) * | 2002-01-04 | 2007-08-23 | Shannon Jones | Method for accessing data via voice |
US20040254791A1 (en) * | 2003-03-01 | 2004-12-16 | Coifman Robert E. | Method and apparatus for improving the transcription accuracy of speech recognition software |
US20040264652A1 (en) * | 2003-06-24 | 2004-12-30 | Erhart George W. | Method and apparatus for validating agreement between textual and spoken representations of words |
US20050010422A1 (en) * | 2003-07-07 | 2005-01-13 | Canon Kabushiki Kaisha | Speech processing apparatus and method |
US20070027693A1 (en) * | 2005-07-27 | 2007-02-01 | Nec Corporation | Voice recognition system and method |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8209171B2 (en) | 2007-08-07 | 2012-06-26 | Aurix Limited | Methods and apparatus relating to searching of spoken audio data |
GB2451938A (en) * | 2007-08-07 | 2009-02-18 | Aurix Ltd | Methods and apparatus for searching of spoken audio data |
US20090043581A1 (en) * | 2007-08-07 | 2009-02-12 | Aurix Limited | Methods and apparatus relating to searching of spoken audio data |
US20090228270A1 (en) * | 2008-03-05 | 2009-09-10 | Microsoft Corporation | Recognizing multiple semantic items from single utterance |
US8725492B2 (en) | 2008-03-05 | 2014-05-13 | Microsoft Corporation | Recognizing multiple semantic items from single utterance |
US8543393B2 (en) | 2008-05-20 | 2013-09-24 | Calabrio, Inc. | Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms |
AU2009249165B2 (en) * | 2008-05-20 | 2012-02-02 | Calabrio, Inc. | Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms |
WO2009143135A1 (en) * | 2008-05-20 | 2009-11-26 | Calabrio, Inc. | Systems and methods of improving automated speech recognition accuracy using statistical analysis of search terms |
US20100161335A1 (en) * | 2008-12-22 | 2010-06-24 | Nortel Networks Limited | Method and system for detecting a relevant utterance |
US8548812B2 (en) * | 2008-12-22 | 2013-10-01 | Avaya Inc. | Method and system for detecting a relevant utterance in a voice session |
US20100235167A1 (en) * | 2009-03-13 | 2010-09-16 | Francois Bourdon | Speech recognition learning system and method |
US8417526B2 (en) | 2009-03-13 | 2013-04-09 | Adacel, Inc. | Speech recognition learning system and method |
US20100312557A1 (en) * | 2009-06-08 | 2010-12-09 | Microsoft Corporation | Progressive application of knowledge sources in multistage speech recognition |
US8386251B2 (en) | 2009-06-08 | 2013-02-26 | Microsoft Corporation | Progressive application of knowledge sources in multistage speech recognition |
US20140330562A1 (en) * | 2011-01-26 | 2014-11-06 | TrackThings LLC | Method and Apparatus for Obtaining Information from the Web |
US9299348B2 (en) * | 2011-01-26 | 2016-03-29 | TrackThings LLC | Method and apparatus for obtaining information from the web |
US20120215538A1 (en) * | 2011-02-17 | 2012-08-23 | Cisco Technology, Inc. | Performance measurement for customer contact centers |
Also Published As
Publication number | Publication date |
---|---|
US7752043B2 (en) | 2010-07-06 |
CA2567230C (en) | 2011-04-05 |
US20120026280A1 (en) | 2012-02-02 |
US9171547B2 (en) | 2015-10-27 |
US20080082329A1 (en) | 2008-04-03 |
CA2567230A1 (en) | 2007-02-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9171547B2 (en) | Multi-pass speech analytics | |
US8005676B2 (en) | Speech analysis using statistical learning | |
US10249304B2 (en) | Method and system for using conversational biometrics and speaker identification/verification to filter voice streams | |
US7991613B2 (en) | Analyzing audio components and generating text with integrated additional session information | |
US11024295B2 (en) | Filtering data in an audio stream | |
US10319366B2 (en) | Predicting recognition quality of a phrase in automatic speech recognition systems | |
US9571652B1 (en) | Enhanced diarization systems, media and methods of use | |
US8798255B2 (en) | Methods and apparatus for deep interaction analysis | |
US8311824B2 (en) | Methods and apparatus for language identification | |
US20150179165A1 (en) | System and method for caller intent labeling of the call-center conversations | |
WO2015019662A1 (en) | Analysis subject determination device and analysis subject determination method | |
US10522135B2 (en) | System and method for segmenting audio files for transcription | |
US20240312466A1 (en) | Systems and Methods for Distinguishing Between Human Speech and Machine Generated Speech | |
EP3641286B1 (en) | Call recording system for automatically storing a call candidate and call recording method | |
US20110197206A1 (en) | System, Method And Program Product For Analyses Based On Agent-Customer Interactions And Concurrent System Activity By Agents | |
US12170082B1 (en) | On-the-fly transcription/redaction of voice-over-IP calls | |
CN113744712A (en) | Intelligent outbound voice splicing method, device, equipment, medium and program product | |
US20250104700A1 (en) | Systems and methods for interaction analytics | |
Guo et al. | IT-Enabled Role Playing in Service Encounter: Design a Customer Emotion Management System in Call Centers | |
CN115567642A (en) | Monitoring method and device for crowdsourcing customer service, computer equipment and storage medium | |
CN119110022A (en) | A method and system for processing cloud phone call data based on big data technology |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CREDIT SUISSE AG, NEW YORK Free format text: SECURITY AGREEMENT;ASSIGNOR:VERINT SYSTEMS INC.;REEL/FRAME:026208/0727 Effective date: 20110429 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: VERINT AMERICAS INC., NEW YORK Free format text: RELEASE OF SECURITY INTEREST IN PATENT RIGHTS;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT;REEL/FRAME:031448/0373 Effective date: 20130918 Owner name: VERINT SYSTEMS INC., NEW YORK Free format text: RELEASE OF SECURITY INTEREST IN PATENT RIGHTS;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT;REEL/FRAME:031448/0373 Effective date: 20130918 Owner name: VERINT VIDEO SOLUTIONS INC., NEW YORK Free format text: RELEASE OF SECURITY INTEREST IN PATENT RIGHTS;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT;REEL/FRAME:031448/0373 Effective date: 20130918 |