US20160019801A1 - System and method for improving presentation skills - Google Patents

System and method for improving presentation skills Download PDF

Info

Publication number
US20160019801A1
US20160019801A1 US14/755,632 US201514755632A US2016019801A1 US 20160019801 A1 US20160019801 A1 US 20160019801A1 US 201514755632 A US201514755632 A US 201514755632A US 2016019801 A1 US2016019801 A1 US 2016019801A1
Authority
US
United States
Prior art keywords
user
presentation
oral
speech
executing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/755,632
Inventor
Danielle A. Feerst
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
AutismSees LLC
Original Assignee
AutismSees LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by AutismSees LLC filed Critical AutismSees LLC
Priority to US14/755,632 priority Critical patent/US20160019801A1/en
Assigned to AutismSees LLC reassignment AutismSees LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FEERST, DANIELLE A.
Publication of US20160019801A1 publication Critical patent/US20160019801A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/06Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
    • G09B5/065Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/113Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining or recording eye movement
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/163Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state by tracking eye movement, gaze, or pupil change
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/165Evaluating the state of mind, e.g. depression, anxiety
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/167Personality evaluation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/168Evaluating attention deficit, hyperactivity
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • G09B17/003Teaching reading electrically operated apparatus or devices
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • G09B17/003Teaching reading electrically operated apparatus or devices
    • G09B17/006Teaching reading electrically operated apparatus or devices with audible presentation of the material to be studied
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • G09B17/02Line indicators or other guides or masks
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B17/00Teaching reading
    • G09B17/04Teaching reading for increasing the rate of reading; Reading rate control
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/04Speaking
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/06Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers
    • G09B7/02Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student

Definitions

  • a system and method provides improved human presentation skills and, more particularly, increases eye contact, speech fluency, presentation skills, and vocabulary of persons with Autism Spectrum Disorders (ASD).
  • ASD Autism Spectrum Disorders
  • apps for speech-making and presentations typically only focus on one aspect, e.g., improving language skills, improving reading skills, or the like.
  • One problem with these apps is that, although individuals with ASD can generally use technology well, they often have difficulty applying lessons learned from the app to real-life situations.
  • Other apps for individuals with ASD are targeted for use by caretakers, rather than by the afflicted individuals themselves.
  • each of the persons has a client device that includes a processing device, memory, a user interface, and a display device.
  • the method includes rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation to measure at least one of a user's pupil movement and a user's gaze direction; displaying, e.g., on a separate display screen and/or an upper portion of the display device, an indicia, e.g., an image of an eye and/or a pair of eyes, which may periodically change; periodically displaying within the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, e.g., using eye tracking software and/or an eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted.
  • the method includes rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of
  • the method may include providing speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text; recording the word-based text during the oral presentation; monitoring user diction and user fluency using the speech-to-text technology; comparing the recorded word-based text with the presentation script to identify any oral delivery mistakes; recording instances of any oral delivery mistakes; alerting the user of any oral delivery mistakes; recording instances when the user did not make eye contact with the indicia when prompted; and/or allowing the user to customize the indicia.
  • the present invention relates to an apparatus for improving social and presentation skills of users.
  • the apparatus may include memory for storing computer readable instructions, which, when executed by a processor, create an instance of an application operating on a mobile device; a user interface; a display device; an eye tracking device; and a processing device for executing some of the stored computer readable instructions.
  • executing may include rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation, using the eye tracking device, to measure a user's pupil movement and/or a user's gaze direction; displaying on an upper portion of the display device an indicia, e.g., an eye and/or a pair of eyes, which can be changed periodically; periodically displaying within the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, using the eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted.
  • executing may further include enabling the user to set a timing of the oral presentation and/or enabling the user to record images, e.g., video images, of the user practicing the oral presentation.
  • the apparatus may also include speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text and executing may further include recording the word-based text during the oral presentation; and monitoring user diction and user fluency using the speech-to-text technology. Executing may also include comparing the recorded word-based text with the presentation script to identify any oral delivery mistakes; recording instances of any oral delivery mistakes; alerting the user of any oral delivery mistakes; using recording instances when the user did not make eye contact with the indicia when prompted; and/or allowing the user to customize the indicia.
  • the system includes a client device and a remote server that is coupled to the client device via a communication network.
  • the client device may include memory for storing computer readable instructions, which, when executed by a processor, create an instance of an application operating on a mobile device; a user interface; a display device; an eye tracking device; and a processing device for executing some of the stored computer readable instructions.
  • executing includes: rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation, e.g., using the eye tracking device, to measure a user's pupil movement and/or a user's gaze direction; displaying, e.g., on an upper portion of the display device an indicia, e.g., an eye and/or a pair of eyes; periodically displaying in the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, using the eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted.
  • a presentation script for an oral presentation to be made by a user
  • monitoring eye movement of the user during the oral presentation e.g., using the eye tracking device, to measure a user's pupil movement and/or a user's gaze direction
  • displaying e.g., on an upper portion of the display device an ind
  • the remote server may include a user interface that is configured and arranged to enable third parties to view and access data created when the processing device executes some of the stored computer readable instructions; memory for storing computer readable instructions; a user interface; a display device; a processing device for executing some of the stored computer readable instructions; and a data storage device for storing the presentation script, speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text, video images of the user practicing the oral presentation, a record of the word-based text during the oral presentation, and/or a record of instances of any oral delivery mistakes.
  • FIG. 1A shows a block diagram of an exemplary system for improving the presentation skills of individuals in accordance with some embodiments of the present invention
  • FIG. 1B shows a block diagram of an exemplary apparatus for improving the presentation skills of individuals in accordance with some embodiments of the present invention
  • FIG. 2 shows a flow diagram of an exemplary method for improving the presentation skills of individuals in accordance with some embodiments of the present invention.
  • FIG. 3 shows an illustrative embodiment of a screen shot on the display device of a user's client device in accordance with an embodiment of the present invention.
  • a method and system are described for aiding self-motivated users to overcome anxiety in a social setting, especially a social setting in which the user is making an oral presentation, e.g., a speech, to an audience or interviewing one-on-one.
  • an oral presentation e.g., a speech
  • the disclosed method and system can be used while the user is making the actual presentation, it may be advantageously employed during user rehearsals, when the user is practicing the speech, to build the confidence of the user.
  • the invention will be described being used in and for the latter scenario, which is to say, during rehearsal for a presentation.
  • Those of ordinary skill in the art can appreciate the applicability of the disclosed steps and elements in the former scenario as well.
  • FIG. 1A and FIG. 1B there are shown, respectively, exemplary embodiments of a system 100 for improving social and presentation skills of persons with ASD, i.e., users, and a client device 10 adapted for use in that system 100 .
  • users may practice using any computer system configuration, including hand-held wireless devices such as mobile or cellular telephones, personal digital assistants (PDAs), tablet computers, smartphones, smartpads, smartwatches, Google® glasses, tablet computers, laptop computers, personal computers, gaming systems, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, computers running under virtualization, and/or any other computing device that is capable of capturing audio and/or video data.
  • PDAs personal digital assistants
  • tablet computers smartphones, smartpads, smartwatches, Google® glasses
  • tablet computers laptop computers
  • gaming systems multiprocessor systems
  • microprocessor-based or programmable consumer electronics minicomputers
  • mainframe computers computers running under virtualization
  • any other computing device that is capable of
  • the data store may be embodied using any computer data store, including but not limited to relational databases, non-relational databases (NoSQL, etc.), flat files, in memory databases, and/or key value stores.
  • Examples of such data stores include the MySQL Database Server or ORACLE Database Server offered by ORACLE Corp. of Redwood Shores, Calif., the PostgreSQL Database Server by the PostgreSQL Global Development Group of Berkeley, Calif., the DB2 Database Server offered by IBM, Mongo DB, Cassandra, and Redis.
  • the invention may be practiced using any computer or processing system 100 that may include a general purpose computing or processing device, i.e., client device 10 , including a processing unit 12 , a system memory 14 , a data storage medium 16 , and a system bus 19 that couples various system components including the system memory 14 to the processing unit 12 .
  • Client devices 10 typically include a variety of computer readable media that can form part of the system memory 14 and be read by the processing unit 12 .
  • computer readable media may include computer storage media and/or communication media.
  • the system memory 14 may include computer storage media in the form of volatile and/or nonvolatile memory, such as read only memory (ROM) and random access memory (RAM).
  • ROM read only memory
  • RAM random access memory
  • BIOS basic input/output system
  • RAM typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 12 .
  • the data or program modules may include an operating system 15 , application programs 11 , other program modules, and program data.
  • the operating system 15 may be or include a variety of operating systems such as Microsoft Windows® operating system, the Unix operating system, the Linux operating system, the Mac OS operating system, Google Android operating
  • the memory 14 may include at least one set of instructions that is either permanently (non-volatile) or temporarily (volatile) stored.
  • the processing unit 12 executes the instructions that are stored in order to process data.
  • the set of instructions may include various instructions that perform a particular task or tasks. Such a set of instructions for performing a particular task may be characterized as a program, software program, software, engine, module, component, mechanism, or tool.
  • the client device 10 may include a plurality of software processing modules stored in the memory 14 as described above and executed on the processing unit 12 in the manner described herein.
  • the program modules may be in the form of any suitable programming language, which is converted to machine language or object code to allow the processor or processing units 12 to read the instructions. That is, written lines of programming code or source code, in a particular programming language, may be converted to machine language using a compiler, assembler, or interpreter.
  • the machine language may be binary coded machine instructions specific to a particular computer.
  • any suitable programming language may be used in accordance with the various embodiments of the invention.
  • the programming language used may include assembly language, Basic, C, C++, CSS, HTML, Java, SQL, Perl, Python, Ruby and/or JavaScript, for example.
  • instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired.
  • An encryption module might be used to encrypt data.
  • files or other data may be decrypted using a suitable decryption module.
  • a user may enter commands and information into the client device 10 through a user interface 18 that includes input devices such as a keyboard and pointing device, commonly referred to as a mouse, trackball or touch pad.
  • input devices such as a keyboard and pointing device, commonly referred to as a mouse, trackball or touch pad.
  • Other input devices may include a microphone, joystick, game pad, satellite dish, scanner, voice recognition device, keyboard, touch screen, toggle switch, pushbutton, or the like.
  • These and other input devices are often connected to the processing unit 12 through a user input interface 18 that is coupled to the system bus 19 , but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB).
  • USB universal serial bus
  • the computing environment may also include other removable/non-removable, volatile/nonvolatile computer storage media 16 .
  • a hard disk drive may read or write to non-removable, nonvolatile magnetic media.
  • a magnetic disk drive may read from or writes to a removable, nonvolatile magnetic disk
  • an optical disk drive may read from or write to a removable, nonvolatile optical disk such as a CD-ROM or other optical media.
  • Other removable/non-removable, volatile/nonvolatile computer storage media 16 that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, Storage Area Networking devices, solid state drives, and the like.
  • the storage media 16 are typically connected to the system bus 19 through a removable or non-removable memory interface.
  • the processing unit 12 that executes commands and instructions may be a general purpose computer, but may utilize any of a wide variety of other technologies including a special purpose computer, a microcomputer, mini-computer, mainframe computer, programmed micro-processor, micro-controller, peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit), ASIC (Application Specific Integrated Circuit), a logic circuit, a digital signal processor, a programmable logic device such as an FPGA (Field Programmable Gate Array), PLD (Programmable Logic Device), PLA (Programmable Logic Array), RFID integrated circuits, smart chip, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention.
  • a programmable logic device such as an FPGA (Field Programmable Gate Array), PLD (Programmable Logic Device), PLA (Programmable Logic Array), RFID integrated circuits, smart chip, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention.
  • One or more monitors or display devices 13 may also be connected to the system bus 19 , e.g., via an interface.
  • the client device 10 may also include other peripheral output devices, which may be connected through an output peripheral interface.
  • the client device 10 implementing the invention may operate in a networked environment using logical connections to one or more remote computers.
  • the remote computers typically including many or all of the elements described above.
  • the processing units 12 and/or memories 14 need not be physically in the same location.
  • the system 100 may also include a general purpose computing or processing device, i.e., server device 20 , including a processing unit 22 , a system memory 24 , a data storage medium, and a system bus.
  • server device 20 including a processing unit 22 , a system memory 24 , a data storage medium, and a system bus.
  • each of the processing units 12 , 22 and each of the memories 14 , 24 used by the system 100 may be in geographically distinct locations and be connected so as to communicate with each other in any suitable manner.
  • each of the processing units 12 , 22 and/or memories 14 , 24 may be composed of different physical pieces of equipment.
  • the devices 10 , 20 that embody the invention may communicate with the user via notifications sent over any protocol that can be transmitted over a packet-switched network or telecommunications (“communication”) network 25 .
  • these may include SMS messages, email (SMTP) messages, instant messages (GChat, AIM, Jabber, etc.), social platform messages (Facebook posts and messages, Twitter direct messages, tweets, retweets, etc.), and mobile push notifications (iOS, Android).
  • the methods and systems 100 described may contain software, middleware, hardware, and any combination thereof connected to, coupled with, and/or in communication with a communication network 25 , e.g., the World Wide Web, the Internet, a local area network (LAN), a wide area network (WAN), and so forth.
  • a communication network 25 e.g., the World Wide Web, the Internet, a local area network (LAN), a wide area network (WAN), and so forth.
  • Computing/processing devices 10 , 20 are capable of communicating with each other via the communication network 25 , and it should be appreciated that the various functionalities of the components may be implemented on any number of devices.
  • a communications network 25 generally connects a client device 10 with a server device 20 , and in the case of peer-to-peer communications, connects two peers.
  • the communication may take place via any media such as standard telephone lines, LAN or WAN links (e.g., T1, T3, 56 kb, X.25), broadband connections (ISDN, Frame Relay, ATM), wireless links (802.11, Bluetooth, 3G, CDMA, etc.), and so on.
  • the communications network 25 may take any form, including but not limited to LAN, WAN, wireless (WiFi, WiMAX), or near field (RFID, Bluetooth).
  • the communications network 25 may use any underlying protocols that can transmit Internet protocols, including but not limited to Ethernet, ATM, VPNs (PPPoE, L2TP, etc.), and encryption (SSL, IPSec, etc.).
  • Examples of software apps that may be used in connection with the system 100 include an eye-tracking app 11 a and/or a speech-to-text app 11 b .
  • the eye-tracking app 11 a provides a device that is configured to track, i.e., to measure the location and the changes of location, the movement of either or each of the user's eyes. Such movement may be used to evaluate whether or not the user made eye contact with a “virtual audience” indicia in response to a prompt appearing in the presentation script. The movement may be measured by a change in distance, e.g., in micrometers, and/or by a change of a point of focus, e.g., in degrees.
  • the speech-to-text app 11 b provides a device that is adapted to make a digital textual record of the user's oral presentation of the presentation script being rehearsed. This textual record may then be compared to the presentation script, e.g., word for word, to identify a word(s) and/or a script portion(s) in the presentation script that the user has difficulty speaking and that a listener would have difficulty understanding, or that the user missed or skipped. These identified words or script portions may then be given greater attention in subsequent rehearsals.
  • Representative, commercially-available speech-to-text apps include the AT&T Speech API from AT&T, the HTML5 Speech API from Mozilla Corporation of Mountain View, Calif., the Text to Speech API from iSpeech, Inc. of Newark, N.J., and the Dragon speech recognition software from Nuance Communications, Inc, of Burlington, Mass.
  • the client device 15 may include an image-recording device 23 , e.g., a digital camera or video recorder, and/or an audio-recording device 17 , e.g., a microphone.
  • the image-recording device 23 may be in electronic communication with the processing unit 12 , data storage medium 14 , and user interface 18 for storing the image data locally, e.g., on the client device 10 , and/or with a communication network interface 21 for storing the image remotely, e.g., at the server device 20 .
  • the image data taken by the digital camera 23 may be uploaded onto a removable memory device, e.g., a memory stick, flash drive, or the like, and subsequently downloaded onto the processing unit that is in electronic communication with a data storage medium, a communication network interface, or the like.
  • the audio-recording device 17 may be in electronic communication with the processing device 12 , data storage medium 14 , and user interface 18 for storing the audio data locally, e.g., on the client device 10 , and/or with a communication network interface 21 for storing the audio data remotely, e.g., at the server device 20 .
  • the audio data recorded by the microphone 17 may be uploaded onto a removable memory device, e.g., a memory stick, flash drive, or the like, and subsequently downloaded onto a processing unit that is in electronic communication with a data storage medium, a communication network interface, and the like.
  • a removable memory device e.g., a memory stick, flash drive, or the like
  • a processing unit that is in electronic communication with a data storage medium, a communication network interface, and the like.
  • a method using the device 10 and system 100 will now be described.
  • the method is not to be construed as being practiced simply by the client device 10 and system 100 described hereinabove. Indeed, there are a myriad of devices having a processing device, memory, a user interface, and a display device that can be programmed or structured and arranged to perform the steps described in greater detail below.
  • the method will be described for persons having ASD, those of ordinary skill in the art can appreciate the applicability of the invention to all persons wishing to improve their social and presentation skills.
  • the user is equipped with a client device having a processing device, memory, a user interface, and a display device.
  • the memory includes computer-readable instructions that are executable by the processing device.
  • the user or a third party on the user's behalf may download a presentation script into a data file stored in the client device (STEP 1 ).
  • the presentation script can be downloaded in a data file in the remote server device, in which case, the remote server device is adapted to upload the presentation onto the user's client device.
  • the user may import the data of the presentation script from, for example, DropboxTM or some other cloud-based data storage that allows members to store and share files and other data.
  • an image-recording device (STEP 2 B) and/or an audio-recording device (STEP 2 A) may be activated.
  • Activation may include recording video and audio data of the user while she is rehearsing or making a presentation and, further, storing the video data and audio data in data storage provided expressly therefor.
  • Data storage may occur locally, i.e., on the client device, or remotely, i.e., on the server device and/or on DropboxTM or other cloud-based data storage.
  • the user or a third party on the user's behalf may activate one or both of the recording devices (STEP 2 A, STEP 2 B).
  • Third parties may include, for the purpose of illustration and not limitation, medical professionals, speech therapists, teachers, parents, guardians, and so forth.
  • the client device and/or the remote server device may render and display, e.g., on the display device of the client device, the downloaded or uploaded presentation script.
  • the display may advantageously include a scrolling feature that can be adjusted to the user's rate of delivery.
  • the user's prosody may be monitored (STEP 3 ) and further converted to a word-based text (STEP 4 ), e.g., using speech-to-text technology.
  • the processing unit may compare, e.g., word for word, the word-based text to the downloaded presentation script (STEP 5 ).
  • the comparison (STEP 5 ) identifies mistakes (STEP 6 ), e.g., clarity, diction, enunciation, fluency, adherence to the script, and the like.
  • any mistakes identified may be identified and recorded (STEP 7 ) for providing feedback to the user.
  • the processing device of the client device may also provide, e.g., transmit, an alert message to the user.
  • the user may focus further attention on and dedicate future rehearsal time to that portion(s) of the presentation, without having to practice the entire presentation.
  • Records of mistakes may also be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • the image-recording device e.g., a digital camera
  • records the user (STEP 2 B) rehearsing or making the oral presentation.
  • an eye tracking and/or monitoring device e.g., eye tracing software
  • the eye tracking and/or monitoring device may monitor the user's eye movement and/or gaze direction (STEP 8 ).
  • the eye tracking and/or monitoring device may measure the user's pupil movement to track the user's progress through the presentation script. Instances in which the user's pupil movement remains fixed at a certain location in the script for a pre-determined period of time may be identified and recorded for providing feedback to the user.
  • the processing unit of the client device may also provide, e.g., transmit, an alert message to the user.
  • Records of lack of eye movement may also be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • the processing unit may integrate a prompt or cue into the presentation script (STEP 9 ).
  • This visual prompt (STEP 9 ) may be provided to cue the user to establish eye contact with her audience.
  • the prompt is meant to cue the user to look at some indicia displayed, for example, on an upper or other portion of the display device or, alternatively, on a separate display device.
  • Illustrative examples of the indicia at which the user is expected to direct her gaze and/or to move her eyes towards can include an image of a human eye, a pair of eyes, a pair of eyes with a human forehead, a face, a group of faces, and so forth.
  • the indicia can change repeatedly, e.g., be looped, so that at each prompt the user's gaze is directed to a new eye or set of eyes.
  • FIG. 3 illustrates an illustrative embodiment of a screen shot showing the user's client device 30 having a display device 32 on which the presentation script 34 is rendered and displayed. Visual prompts 35 appear at discrete locations of the presentation script 34 . Indicia, in this case a pair of eyes 33 , appear at the top of the display device 32 .
  • the user's eye movement may be measured, e.g., using the eye tracking and monitoring device, at occurrence of the visual prompt.
  • the measurement device is sufficiently accurate to determine whether or not the user has made eye contact (STEP 10 ) with her “virtual audience,” i.e., the pair of eyes indicia.
  • any missed opportunities for making prompted eye contact (“misses”) may be identified and recorded (STEP 11 ) for providing feedback to the user.
  • the processing device of the client device may also provide, e.g., transmit, an alert message to the user. Records of eye contact misses may advantageously be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • Additional features of the method may include enabling the user to time the oral presentation, e.g., using a timing device in communication with the processing unit of the client device and/or enabling the user to customize her “virtual audience.”
  • users may be able to make or import their own images for use as indicia, to use images of people with whom they are more familiar, e.g., parents, siblings, and the like. Users may also be able to turn off the “virtual audience” feature.
  • the method may also include a reward feature by which the user or a third party may establish a goal for a particular rehearsal and/or rewards may be based on the number of times the user rehearses the oral presentation or improves in her performance in the oral presentation.
  • Audio and video data captured during a rehearsal may also be exported to an online video website, e.g., YouTube.
  • Additional uses of the method and system described herein may facilitate providing the user with experience involving open-ended dialog exchange, such as at a job interview or meeting new individuals at a social event.

Abstract

A system and method for improving social and presentation skills of persons with a social communication disorder such as Autism. Social anxiety or a lack of confidence, the method including: rendering and displaying on a display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation to measure a user's pupil movement and/or a user's gaze direction; displaying an indicia on a separate display screen and/or an upper portion of the display device; periodically displaying within the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement using eye tracking software and/or an eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims priority to U.S. Provisional Patent Application No. 61/833,303 filed on Jun. 10, 2013, which is incorporated by reference as if set forth herein in its entirety.
  • FIELD OF THE INVENTION
  • A system and method provides improved human presentation skills and, more particularly, increases eye contact, speech fluency, presentation skills, and vocabulary of persons with Autism Spectrum Disorders (ASD).
  • BACKGROUND OF THE INVENTION
  • Individuals, especially young adults and older teens afflicted with ASD, often have problems communicating with other individuals, especially with audiences or groups of people. For many individuals with ASD, the consequences of the disorder may manifest as, inter alia, difficulties in making eye-to-eye contact, in presenting ideas fluently, and in socializing and engaging with others. The short- and long-term effect of these social skills deficits may hinder the individual's relationship development, academic success, and professional advancement.
  • Some commercially-available software applications (“apps”) for speech-making and presentations typically only focus on one aspect, e.g., improving language skills, improving reading skills, or the like. One problem with these apps is that, although individuals with ASD can generally use technology well, they often have difficulty applying lessons learned from the app to real-life situations. Other apps for individuals with ASD are targeted for use by caretakers, rather than by the afflicted individuals themselves.
  • BRIEF SUMMARY OF THE INVENTION
  • Accordingly, it is desirable to provide a system and method for improving presentation skills, especially skills of individuals with social skill and communication disorders such as Autism. More specifically, it is desirable to provide a system and method for improving presentation skills by targeting more than one aspect of making a presentation, i.e., delivering a speech to an audience, with particular focus on improving eye contact skills, especially among higher functioning individuals who recognize their own social anxiety and desire self-improvement in that area. Moreover, it is desirable to provide a user-friendly system and method that enables higher-functioning individuals with ASD to overcome the anxiety of making eye contact with an audience of one or multiple individuals by practicing the presentation, e.g., speech, using visual prompts in the text of the script of the presentation.
  • It is particularly desirable to integrate technology and/or apps for tracking the eye movement of a user over time, especially at discrete times within the presentation script. It is also desirable to enable system and method users to record, e.g., make a video, and to time themselves making the presentation while rehearsing the speech. Also desirable is to include speech-to-text technology and/or apps to monitor the diction and fluency of the user during rehearsals, to identify any problems with, for example, enunciation, elision, fluency, and adherence to the presentation script.
  • In a first aspect of a method for improving social and presentation skills of persons with an Autism Spectrum Disorder, each of the persons has a client device that includes a processing device, memory, a user interface, and a display device. In some embodiments, the method includes rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation to measure at least one of a user's pupil movement and a user's gaze direction; displaying, e.g., on a separate display screen and/or an upper portion of the display device, an indicia, e.g., an image of an eye and/or a pair of eyes, which may periodically change; periodically displaying within the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, e.g., using eye tracking software and/or an eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted. In variations of the embodiment, the method may further include enabling the user to set a timing of the oral presentation and/or enabling the user to record images, e.g., video images, of the user practicing the oral presentation.
  • In further variations, the method may include providing speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text; recording the word-based text during the oral presentation; monitoring user diction and user fluency using the speech-to-text technology; comparing the recorded word-based text with the presentation script to identify any oral delivery mistakes; recording instances of any oral delivery mistakes; alerting the user of any oral delivery mistakes; recording instances when the user did not make eye contact with the indicia when prompted; and/or allowing the user to customize the indicia.
  • In a second aspect, the present invention relates to an apparatus for improving social and presentation skills of users. In some embodiments, the apparatus may include memory for storing computer readable instructions, which, when executed by a processor, create an instance of an application operating on a mobile device; a user interface; a display device; an eye tracking device; and a processing device for executing some of the stored computer readable instructions. In some variations of the embodiment, executing may include rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation, using the eye tracking device, to measure a user's pupil movement and/or a user's gaze direction; displaying on an upper portion of the display device an indicia, e.g., an eye and/or a pair of eyes, which can be changed periodically; periodically displaying within the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, using the eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted. In variations, executing may further include enabling the user to set a timing of the oral presentation and/or enabling the user to record images, e.g., video images, of the user practicing the oral presentation.
  • In further variations, the apparatus may also include speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text and executing may further include recording the word-based text during the oral presentation; and monitoring user diction and user fluency using the speech-to-text technology. Executing may also include comparing the recorded word-based text with the presentation script to identify any oral delivery mistakes; recording instances of any oral delivery mistakes; alerting the user of any oral delivery mistakes; using recording instances when the user did not make eye contact with the indicia when prompted; and/or allowing the user to customize the indicia.
  • In a third aspect, a system for improving social and presentation skills of persons with an autism spectrum disorder is disclosed. In some embodiments, the system includes a client device and a remote server that is coupled to the client device via a communication network. In variations of the embodiment, the client device may include memory for storing computer readable instructions, which, when executed by a processor, create an instance of an application operating on a mobile device; a user interface; a display device; an eye tracking device; and a processing device for executing some of the stored computer readable instructions. In some implementations, executing includes: rendering and displaying on the display device a presentation script for an oral presentation to be made by a user; monitoring eye movement of the user during the oral presentation, e.g., using the eye tracking device, to measure a user's pupil movement and/or a user's gaze direction; displaying, e.g., on an upper portion of the display device an indicia, e.g., an eye and/or a pair of eyes; periodically displaying in the presentation script a visual prompt to cue the user to look at the indicia; measuring eye movement, using the eye tracking device, at occurrence of the visual prompt; and evaluating whether the user made eye contact with the indicia when prompted. In other implementations, the remote server may include a user interface that is configured and arranged to enable third parties to view and access data created when the processing device executes some of the stored computer readable instructions; memory for storing computer readable instructions; a user interface; a display device; a processing device for executing some of the stored computer readable instructions; and a data storage device for storing the presentation script, speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text, video images of the user practicing the oral presentation, a record of the word-based text during the oral presentation, and/or a record of instances of any oral delivery mistakes.
  • Although the invention is disclosed and described in the context of users who are individuals with ASD, those of ordinary skill in the art can appreciate that the system and method described herein have applicability to all individuals who make presentations to audiences large and small, to all individuals who may have some degree of social anxiety, and to any individuals desiring to hone their oral presentation skills, especially with respect to making improvements to making eye contact with one's audience. Accordingly, the disclosure is not intended to be read narrowly or to be construed as being limited just to that application of the technology described.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings are not intended to be drawn to scale. In the drawings, each identical or similar component that is illustrated in various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
  • FIG. 1A shows a block diagram of an exemplary system for improving the presentation skills of individuals in accordance with some embodiments of the present invention;
  • FIG. 1B shows a block diagram of an exemplary apparatus for improving the presentation skills of individuals in accordance with some embodiments of the present invention;
  • FIG. 2 shows a flow diagram of an exemplary method for improving the presentation skills of individuals in accordance with some embodiments of the present invention; and
  • FIG. 3 shows an illustrative embodiment of a screen shot on the display device of a user's client device in accordance with an embodiment of the present invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • A method and system are described for aiding self-motivated users to overcome anxiety in a social setting, especially a social setting in which the user is making an oral presentation, e.g., a speech, to an audience or interviewing one-on-one. Although the disclosed method and system can be used while the user is making the actual presentation, it may be advantageously employed during user rehearsals, when the user is practicing the speech, to build the confidence of the user. For simplicity, the invention will be described being used in and for the latter scenario, which is to say, during rehearsal for a presentation. Those of ordinary skill in the art can appreciate the applicability of the disclosed steps and elements in the former scenario as well.
  • System
  • Referring to FIG. 1A and FIG. 1B, there are shown, respectively, exemplary embodiments of a system 100 for improving social and presentation skills of persons with ASD, i.e., users, and a client device 10 adapted for use in that system 100. Preferably, users may practice using any computer system configuration, including hand-held wireless devices such as mobile or cellular telephones, personal digital assistants (PDAs), tablet computers, smartphones, smartpads, smartwatches, Google® glasses, tablet computers, laptop computers, personal computers, gaming systems, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, computers running under virtualization, and/or any other computing device that is capable of capturing audio and/or video data.
  • The data store may be embodied using any computer data store, including but not limited to relational databases, non-relational databases (NoSQL, etc.), flat files, in memory databases, and/or key value stores. Examples of such data stores include the MySQL Database Server or ORACLE Database Server offered by ORACLE Corp. of Redwood Shores, Calif., the PostgreSQL Database Server by the PostgreSQL Global Development Group of Berkeley, Calif., the DB2 Database Server offered by IBM, Mongo DB, Cassandra, and Redis.
  • The invention may be practiced using any computer or processing system 100 that may include a general purpose computing or processing device, i.e., client device 10, including a processing unit 12, a system memory 14, a data storage medium 16, and a system bus 19 that couples various system components including the system memory 14 to the processing unit 12.
  • Client devices 10 typically include a variety of computer readable media that can form part of the system memory 14 and be read by the processing unit 12. By way of example, and not limitation, computer readable media may include computer storage media and/or communication media. The system memory 14 may include computer storage media in the form of volatile and/or nonvolatile memory, such as read only memory (ROM) and random access memory (RAM). A basic input/output system (BIOS), containing the basic routines that help to transfer information between components, such as during start-up, is typically stored in ROM. RAM typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 12. The data or program modules may include an operating system 15, application programs 11, other program modules, and program data. The operating system 15 may be or include a variety of operating systems such as Microsoft Windows® operating system, the Unix operating system, the Linux operating system, the Mac OS operating system, Google Android operating system, Apple iOS operating system, or another operating system or platform.
  • At a minimum, the memory 14 may include at least one set of instructions that is either permanently (non-volatile) or temporarily (volatile) stored. The processing unit 12 executes the instructions that are stored in order to process data. The set of instructions may include various instructions that perform a particular task or tasks. Such a set of instructions for performing a particular task may be characterized as a program, software program, software, engine, module, component, mechanism, or tool.
  • The client device 10 may include a plurality of software processing modules stored in the memory 14 as described above and executed on the processing unit 12 in the manner described herein. The program modules may be in the form of any suitable programming language, which is converted to machine language or object code to allow the processor or processing units 12 to read the instructions. That is, written lines of programming code or source code, in a particular programming language, may be converted to machine language using a compiler, assembler, or interpreter. The machine language may be binary coded machine instructions specific to a particular computer.
  • Any suitable programming language may be used in accordance with the various embodiments of the invention. Illustratively, the programming language used may include assembly language, Basic, C, C++, CSS, HTML, Java, SQL, Perl, Python, Ruby and/or JavaScript, for example. Further, it is not necessary that a single type of instruction or programming language be utilized in conjunction with the operation of the system and method of the invention. Rather, any number of different programming languages may be utilized as is necessary or desirable.
  • Also, the instructions and/or data used in the practice of the invention may utilize any compression or encryption technique or algorithm, as may be desired. An encryption module might be used to encrypt data. Further, files or other data may be decrypted using a suitable decryption module.
  • A user may enter commands and information into the client device 10 through a user interface 18 that includes input devices such as a keyboard and pointing device, commonly referred to as a mouse, trackball or touch pad. Other input devices may include a microphone, joystick, game pad, satellite dish, scanner, voice recognition device, keyboard, touch screen, toggle switch, pushbutton, or the like. These and other input devices are often connected to the processing unit 12 through a user input interface 18 that is coupled to the system bus 19, but may be connected by other interface and bus structures, such as a parallel port, game port or a universal serial bus (USB).
  • The computing environment may also include other removable/non-removable, volatile/nonvolatile computer storage media 16. For example, a hard disk drive may read or write to non-removable, nonvolatile magnetic media. A magnetic disk drive may read from or writes to a removable, nonvolatile magnetic disk, and an optical disk drive may read from or write to a removable, nonvolatile optical disk such as a CD-ROM or other optical media. Other removable/non-removable, volatile/nonvolatile computer storage media 16 that can be used in the exemplary operating environment include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, Storage Area Networking devices, solid state drives, and the like. The storage media 16 are typically connected to the system bus 19 through a removable or non-removable memory interface.
  • The processing unit 12 that executes commands and instructions may be a general purpose computer, but may utilize any of a wide variety of other technologies including a special purpose computer, a microcomputer, mini-computer, mainframe computer, programmed micro-processor, micro-controller, peripheral integrated circuit element, a CSIC (Customer Specific Integrated Circuit), ASIC (Application Specific Integrated Circuit), a logic circuit, a digital signal processor, a programmable logic device such as an FPGA (Field Programmable Gate Array), PLD (Programmable Logic Device), PLA (Programmable Logic Array), RFID integrated circuits, smart chip, or any other device or arrangement of devices that is capable of implementing the steps of the processes of the invention.
  • One or more monitors or display devices 13 may also be connected to the system bus 19, e.g., via an interface. In addition to display devices 13, the client device 10 may also include other peripheral output devices, which may be connected through an output peripheral interface. The client device 10 implementing the invention may operate in a networked environment using logical connections to one or more remote computers. The remote computers typically including many or all of the elements described above.
  • It should be appreciated that the processing units 12 and/or memories 14 need not be physically in the same location. For example, in some implementations, the system 100 may also include a general purpose computing or processing device, i.e., server device 20, including a processing unit 22, a system memory 24, a data storage medium, and a system bus. Hence, each of the processing units 12, 22 and each of the memories 14, 24 used by the system 100 may be in geographically distinct locations and be connected so as to communicate with each other in any suitable manner. Additionally, it is appreciated that each of the processing units 12, 22 and/or memories 14, 24 may be composed of different physical pieces of equipment.
  • The devices 10, 20 that embody the invention may communicate with the user via notifications sent over any protocol that can be transmitted over a packet-switched network or telecommunications (“communication”) network 25. By way of example, and not limitation, these may include SMS messages, email (SMTP) messages, instant messages (GChat, AIM, Jabber, etc.), social platform messages (Facebook posts and messages, Twitter direct messages, tweets, retweets, etc.), and mobile push notifications (iOS, Android).
  • It is understood that the methods and systems 100 described may contain software, middleware, hardware, and any combination thereof connected to, coupled with, and/or in communication with a communication network 25, e.g., the World Wide Web, the Internet, a local area network (LAN), a wide area network (WAN), and so forth. Computing/ processing devices 10, 20 are capable of communicating with each other via the communication network 25, and it should be appreciated that the various functionalities of the components may be implemented on any number of devices.
  • The invention may be practiced using any communications network 25 capable of transmitting Internet protocols. A communications network 25 generally connects a client device 10 with a server device 20, and in the case of peer-to-peer communications, connects two peers. The communication may take place via any media such as standard telephone lines, LAN or WAN links (e.g., T1, T3, 56 kb, X.25), broadband connections (ISDN, Frame Relay, ATM), wireless links (802.11, Bluetooth, 3G, CDMA, etc.), and so on. The communications network 25 may take any form, including but not limited to LAN, WAN, wireless (WiFi, WiMAX), or near field (RFID, Bluetooth). The communications network 25 may use any underlying protocols that can transmit Internet protocols, including but not limited to Ethernet, ATM, VPNs (PPPoE, L2TP, etc.), and encryption (SSL, IPSec, etc.).
  • Examples of software apps that may be used in connection with the system 100 include an eye-tracking app 11 a and/or a speech-to-text app 11 b. The eye-tracking app 11 a provides a device that is configured to track, i.e., to measure the location and the changes of location, the movement of either or each of the user's eyes. Such movement may be used to evaluate whether or not the user made eye contact with a “virtual audience” indicia in response to a prompt appearing in the presentation script. The movement may be measured by a change in distance, e.g., in micrometers, and/or by a change of a point of focus, e.g., in degrees. Representative, commercially-available eye tracking apps include EyeWorks™ from EyeTracking, Inc. of San Diego, Calif. and S2 Eye Tracker from Mirametrix, Inc. of Montreal, Canada. The speech-to-text app 11 b provides a device that is adapted to make a digital textual record of the user's oral presentation of the presentation script being rehearsed. This textual record may then be compared to the presentation script, e.g., word for word, to identify a word(s) and/or a script portion(s) in the presentation script that the user has difficulty speaking and that a listener would have difficulty understanding, or that the user missed or skipped. These identified words or script portions may then be given greater attention in subsequent rehearsals. Representative, commercially-available speech-to-text apps include the AT&T Speech API from AT&T, the HTML5 Speech API from Mozilla Corporation of Mountain View, Calif., the Text to Speech API from iSpeech, Inc. of Newark, N.J., and the Dragon speech recognition software from Nuance Communications, Inc, of Burlington, Mass.
  • In some variations, the client device 15 may include an image-recording device 23, e.g., a digital camera or video recorder, and/or an audio-recording device 17, e.g., a microphone. In some variations, the image-recording device 23 may be in electronic communication with the processing unit 12, data storage medium 14, and user interface 18 for storing the image data locally, e.g., on the client device 10, and/or with a communication network interface 21 for storing the image remotely, e.g., at the server device 20. Alternatively, the image data taken by the digital camera 23 may be uploaded onto a removable memory device, e.g., a memory stick, flash drive, or the like, and subsequently downloaded onto the processing unit that is in electronic communication with a data storage medium, a communication network interface, or the like. Similarly, the audio-recording device 17 may be in electronic communication with the processing device 12, data storage medium 14, and user interface 18 for storing the audio data locally, e.g., on the client device 10, and/or with a communication network interface 21 for storing the audio data remotely, e.g., at the server device 20. Alternatively, the audio data recorded by the microphone 17 may be uploaded onto a removable memory device, e.g., a memory stick, flash drive, or the like, and subsequently downloaded onto a processing unit that is in electronic communication with a data storage medium, a communication network interface, and the like.
  • Method
  • Having described a client device 10 and a system 100 for improving social and presentation skills in persons with ASD, a method using the device 10 and system 100 will now be described. Those of ordinary skill in the art can appreciate that the method is not to be construed as being practiced simply by the client device 10 and system 100 described hereinabove. Indeed, there are a myriad of devices having a processing device, memory, a user interface, and a display device that can be programmed or structured and arranged to perform the steps described in greater detail below. Moreover, although the method will be described for persons having ASD, those of ordinary skill in the art can appreciate the applicability of the invention to all persons wishing to improve their social and presentation skills.
  • Referring to FIG. 2, there is shown a flow chart of one exemplary embodiment of a method for improving social and presentation skills of persons with ASD. In some embodiments, the user is equipped with a client device having a processing device, memory, a user interface, and a display device. Preferably, the memory includes computer-readable instructions that are executable by the processing device. In a first step, prior to rehearsing and/or making a presentation, the user or a third party on the user's behalf may download a presentation script into a data file stored in the client device (STEP 1). Alternatively, the presentation script can be downloaded in a data file in the remote server device, in which case, the remote server device is adapted to upload the presentation onto the user's client device. In another alternative, the user may import the data of the presentation script from, for example, Dropbox™ or some other cloud-based data storage that allows members to store and share files and other data.
  • Contemporaneously with rehearsing and/or making a presentation, an image-recording device (STEP 2B) and/or an audio-recording device (STEP 2A) may be activated. Activation may include recording video and audio data of the user while she is rehearsing or making a presentation and, further, storing the video data and audio data in data storage provided expressly therefor. Data storage may occur locally, i.e., on the client device, or remotely, i.e., on the server device and/or on Dropbox™ or other cloud-based data storage. Once again, the user or a third party on the user's behalf may activate one or both of the recording devices (STEP 2A, STEP 2B). Third parties may include, for the purpose of illustration and not limitation, medical professionals, speech therapists, teachers, parents, guardians, and so forth.
  • With the image and audio-recording devices recording image and storing video data and audio data, the client device and/or the remote server device may render and display, e.g., on the display device of the client device, the downloaded or uploaded presentation script. The display may advantageously include a scrolling feature that can be adjusted to the user's rate of delivery. As the user “speaks the speech,” in addition to being recorded, the user's prosody may be monitored (STEP 3) and further converted to a word-based text (STEP 4), e.g., using speech-to-text technology. As the speech-to-text technology converts the user's speech to word-based text (STEP 4), the processing unit may compare, e.g., word for word, the word-based text to the downloaded presentation script (STEP 5). The comparison (STEP 5) identifies mistakes (STEP 6), e.g., clarity, diction, enunciation, fluency, adherence to the script, and the like.
  • Advantageously, any mistakes identified (STEP 6) may be identified and recorded (STEP 7) for providing feedback to the user. In some implementations, the processing device of the client device may also provide, e.g., transmit, an alert message to the user. By recording individual oral delivery mistakes, the user may focus further attention on and dedicate future rehearsal time to that portion(s) of the presentation, without having to practice the entire presentation. Records of mistakes may also be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • Also, while the user is rehearsing or making an oral presentation, the image-recording device, e.g., a digital camera, records the user (STEP 2B) rehearsing or making the oral presentation. Contemporaneous with recording video image data, an eye tracking and/or monitoring device, e.g., eye tracing software, may monitor the user's eye movement and/or gaze direction (STEP 8). For example, the eye tracking and/or monitoring device may measure the user's pupil movement to track the user's progress through the presentation script. Instances in which the user's pupil movement remains fixed at a certain location in the script for a pre-determined period of time may be identified and recorded for providing feedback to the user. In some implementations, the processing unit of the client device may also provide, e.g., transmit, an alert message to the user. Records of lack of eye movement may also be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • Advantageously, periodically during the oral rehearsal or presentation, the processing unit may integrate a prompt or cue into the presentation script (STEP 9). This visual prompt (STEP 9) may be provided to cue the user to establish eye contact with her audience. During rehearsal of an oral presentation, in which the presence of an audience is not likely, the prompt is meant to cue the user to look at some indicia displayed, for example, on an upper or other portion of the display device or, alternatively, on a separate display device. Illustrative examples of the indicia at which the user is expected to direct her gaze and/or to move her eyes towards can include an image of a human eye, a pair of eyes, a pair of eyes with a human forehead, a face, a group of faces, and so forth. In some implementations, the indicia can change repeatedly, e.g., be looped, so that at each prompt the user's gaze is directed to a new eye or set of eyes. FIG. 3 illustrates an illustrative embodiment of a screen shot showing the user's client device 30 having a display device 32 on which the presentation script 34 is rendered and displayed. Visual prompts 35 appear at discrete locations of the presentation script 34. Indicia, in this case a pair of eyes 33, appear at the top of the display device 32.
  • To evaluate whether or not the user has made eye contact with the “visual audience,” i.e., the pair of eyes indicia, when prompted, the user's eye movement may be measured, e.g., using the eye tracking and monitoring device, at occurrence of the visual prompt. The measurement device is sufficiently accurate to determine whether or not the user has made eye contact (STEP 10) with her “virtual audience,” i.e., the pair of eyes indicia. Advantageously, any missed opportunities for making prompted eye contact (“misses”) may be identified and recorded (STEP 11) for providing feedback to the user. In some implementations, the processing device of the client device may also provide, e.g., transmit, an alert message to the user. Records of eye contact misses may advantageously be accessed by or made accessible to third parties, e.g., speech therapists, teachers, and the like, to enable the third parties to work with the user in a targeted, constructive manner.
  • Additional features of the method may include enabling the user to time the oral presentation, e.g., using a timing device in communication with the processing unit of the client device and/or enabling the user to customize her “virtual audience.” For example, in some implementations, users may be able to make or import their own images for use as indicia, to use images of people with whom they are more familiar, e.g., parents, siblings, and the like. Users may also be able to turn off the “virtual audience” feature. The method may also include a reward feature by which the user or a third party may establish a goal for a particular rehearsal and/or rewards may be based on the number of times the user rehearses the oral presentation or improves in her performance in the oral presentation. If the user accomplishes the goal or performs a certain number of rehearsals, she may receive a reward such as an iTune download, a Kiip reward, a discount coupon, and so forth. Audio and video data captured during a rehearsal may also be exported to an online video website, e.g., YouTube.
  • Additional uses of the method and system described herein may facilitate providing the user with experience involving open-ended dialog exchange, such as at a job interview or meeting new individuals at a social event.
  • Thus, the foregoing discussion discloses and describes merely exemplary embodiments of the present invention. As will be understood by those skilled in the art, the present invention may be embodied in other specific forms without departing from the spirit or essential characteristics thereof. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting of the scope of the invention, as well as other claims. The disclosure, including any readily discernible variants of the teachings herein, defines, in part, the scope of the foregoing claim terminology.

Claims (18)

1-23. (canceled)
24. A method for improving social and presentation skills of a person, the person having a client device including a processing device, memory, a user interface, and a display device, the method comprising:
rendering and displaying on the display device a scrolling presentation speech for an oral presentation to be made by a user;
converting words spoken by the user during the oral presentation to a word-based text;
comparing the word-based text with the presentation speech to identify any oral delivery mistakes; and
identifying instances of any oral delivery mistakes.
25. The method of claim 24 further comprising setting a timing of the oral presentation.
26. The method of claim 24 further comprising recording images of the user practicing the oral presentation.
27. The method of claim 26 further comprising recording video images of the user practicing the oral presentation.
28. The method of claim 24 further comprising monitoring at least one of user diction and user fluency.
29. The method of claim 28 further comprising alerting the user of any oral delivery mistakes.
30. The method of claim 24 further comprising enabling at least one third party to view and access the instances of oral delivery mistakes.
31. An apparatus for improving social and presentation skills of persons, the apparatus comprising:
memory for storing computer readable instructions, which, when executed by a processing unit, create an instance of an application operating on a mobile device;
a user interface;
a display device; and
a processing unit for executing some of the stored computer readable instructions, wherein executing comprises:
rendering and displaying on the display device a scrolling presentation speech for an oral presentation to be made by a user;
converting words spoken by the user during the oral presentation to a word-based text, wherein executing further comprises:
comparing the word-based text with the presentation speech to identify any oral delivery mistakes; and
identifying instances of any oral delivery mistakes.
32. The apparatus of claim 31, wherein executing further comprises recording images of the user practicing the oral presentation.
33. The apparatus of claim 32, wherein executing further comprises recording video images of the user practicing the oral presentation.
34. The apparatus of claim 31 further comprising monitoring at least one of user diction and user fluency.
35. The apparatus of claim 31, wherein executing further comprises alerting the user of any oral delivery mistakes.
36. The apparatus of claim 31, wherein executing further comprises enabling at least one third party to view and access the instances of oral delivery mistakes.
37. A system for improving social and presentation skills of persons, the system comprising:
a client device further comprising:
memory for storing computer readable instructions, which, when executed by a processing unit, create an instance of an application operating on a mobile device;
a user interface;
a display device; and
a processing unit for executing some of the stored computer readable instructions, wherein executing comprises:
rendering and displaying on the display device a scrolling presentation speech for an oral presentation to be made by a user,
converting words spoken by the user during the oral presentation to a word-based text,
comparing the word-based text with the presentation speech to identify any oral delivery mistakes, and
identifying instances of any oral delivery mistakes; and
a remote server that is coupled to the client device via a communication network.
38. The system of claim 37, wherein the remote server comprises a user interface that is configured and arranged to enable third parties to view and access data created when the processing unit executes at least some of the stored computer readable instructions.
39. The system of claim 37, wherein the remote server comprises:
memory for storing computer readable instructions;
a user interface;
a display device;
a processing unit for executing at least some of the stored computer readable instructions; and
a data storage device for storing at least one of the presentation speech, speech-to-text technology to convert words spoken by the user during the oral presentation to a word-based text, video images of the user practicing the oral presentation, a record of the word-based text during the oral presentation, and a record of instances of any oral delivery mistakes.
40. The system of claim 37, wherein executing further comprises enabling at least one third party to view and access the instances of oral delivery mistakes.
US14/755,632 2013-06-10 2015-06-30 System and method for improving presentation skills Abandoned US20160019801A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/755,632 US20160019801A1 (en) 2013-06-10 2015-06-30 System and method for improving presentation skills

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201361833303P 2013-06-10 2013-06-10
US14/300,653 US9072478B1 (en) 2013-06-10 2014-06-10 System and method for improving presentation skills
US14/755,632 US20160019801A1 (en) 2013-06-10 2015-06-30 System and method for improving presentation skills

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/300,653 Continuation US9072478B1 (en) 2013-06-10 2014-06-10 System and method for improving presentation skills

Publications (1)

Publication Number Publication Date
US20160019801A1 true US20160019801A1 (en) 2016-01-21

Family

ID=53491838

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/300,653 Expired - Fee Related US9072478B1 (en) 2013-06-10 2014-06-10 System and method for improving presentation skills
US14/755,632 Abandoned US20160019801A1 (en) 2013-06-10 2015-06-30 System and method for improving presentation skills

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/300,653 Expired - Fee Related US9072478B1 (en) 2013-06-10 2014-06-10 System and method for improving presentation skills

Country Status (1)

Country Link
US (2) US9072478B1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10732784B2 (en) 2016-09-01 2020-08-04 University Of Massachusetts System and methods for cuing visual attention
EP4018647A4 (en) * 2019-08-23 2023-08-16 Your Speech Factory AB Electronic device and method for eye-contact training

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9072478B1 (en) * 2013-06-10 2015-07-07 AutismSees LLC System and method for improving presentation skills
US10825353B2 (en) * 2013-08-13 2020-11-03 The Children's Hospital Of Philadelphia Device for enhancement of language processing in autism spectrum disorders through modifying the auditory stream including an acoustic stimulus to reduce an acoustic detail characteristic while preserving a lexicality of the acoustics stimulus
IL263655B2 (en) * 2016-06-14 2023-03-01 Netzer Omry Automatic speech recognition
WO2019207573A1 (en) * 2018-04-25 2019-10-31 Ninispeech Ltd. Diagnosis and treatment of speech and language pathologies by speech to text and natural language processing
CN110349474A (en) * 2019-06-17 2019-10-18 厦门盈趣科技股份有限公司 Local area network communication and the intelligent pointer and teaching method of wireless communication control
US11633668B2 (en) * 2020-10-24 2023-04-25 Motorola Mobility Llc Eye contact prompting communication device
US20220141266A1 (en) * 2020-11-02 2022-05-05 Virtual Sapiens Inc. System and method to improve video conferencing using presence metrics

Citations (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5920838A (en) * 1997-06-02 1999-07-06 Carnegie Mellon University Reading and pronunciation tutor
US6045515A (en) * 1997-04-07 2000-04-04 Lawton; Teri A. Methods and apparatus for diagnosing and remediating reading disorders
US20010046659A1 (en) * 2000-05-16 2001-11-29 William Oster System for improving reading & speaking
US20020099305A1 (en) * 2000-12-28 2002-07-25 Matsushita Electic Works, Ltd. Non-invasive brain function examination
US6704699B2 (en) * 2000-09-05 2004-03-09 Einat H. Nir Language acquisition aide
US20050119894A1 (en) * 2003-10-20 2005-06-02 Cutler Ann R. System and process for feedback speech instruction
US6931587B1 (en) * 1998-01-29 2005-08-16 Philip R. Krause Teleprompter device
US20050273017A1 (en) * 2004-03-26 2005-12-08 Evian Gordon Collective brain measurement system and method
US20060093998A1 (en) * 2003-03-21 2006-05-04 Roel Vertegaal Method and apparatus for communication between humans and devices
US20060270945A1 (en) * 2004-02-11 2006-11-30 Jamshid Ghajar Cognition and motor timing diagnosis using smooth eye pursuit analysis
US7149690B2 (en) * 1999-09-09 2006-12-12 Lucent Technologies Inc. Method and apparatus for interactive language instruction
US20070050151A1 (en) * 2005-08-10 2007-03-01 Shinji Satoh Psychotic manifestation and mental state evaluation apparatus and evaluation method
US7211050B1 (en) * 1999-06-15 2007-05-01 Dimitri Caplygin System for enhancement of neurophysiological processes
US20070166676A1 (en) * 2005-12-15 2007-07-19 Posit Science Corporation Cognitive training using guided eye movements
US7254531B2 (en) * 2000-09-05 2007-08-07 Nir Einat H In-context analysis and automatic translation
US20070218432A1 (en) * 2006-03-15 2007-09-20 Glass Andrew B System and Method for Controlling the Presentation of Material and Operation of External Devices
US20070248938A1 (en) * 2006-01-27 2007-10-25 Rocketreader Pty Ltd Method for teaching reading using systematic and adaptive word recognition training and system for realizing this method.
US20080309889A1 (en) * 2007-06-18 2008-12-18 Keynote Products Llc Speech Prompter Apparatus
US20090051877A1 (en) * 2006-12-15 2009-02-26 Posit Science Corporation Cognitive training using guided eye movements
US20090089062A1 (en) * 2007-10-01 2009-04-02 Fang Lu Public speaking self-evaluation tool
US7678047B2 (en) * 2001-11-13 2010-03-16 Electronic Navigation Research Institute Chaologic brain function diagnosis apparatus
US7697825B2 (en) * 2004-08-18 2010-04-13 Sunplus Technology Co., Ltd. DVD player with language learning function
US20100092929A1 (en) * 2008-10-14 2010-04-15 Ohio University Cognitive and Linguistic Assessment Using Eye Tracking
US7815507B2 (en) * 2004-06-18 2010-10-19 Igt Game machine user interface using a non-contact eye motion recognition device
US20110026779A1 (en) * 2008-12-24 2011-02-03 David Matsumoto Systems and methods for analyzing facial expressions, identifying intent and transforming images through review of facial expressions
US20110063571A1 (en) * 2009-09-16 2011-03-17 Duffy Charles J Method and system for quantitative assessment of visual contrast sensitivity
US20110065069A1 (en) * 2009-09-16 2011-03-17 Duffy Charles J Method and system for quantitative assessment of verbal recognition memory
US20110123967A1 (en) * 2009-11-24 2011-05-26 Xerox Corporation Dialog system for comprehension evaluation
US7972278B2 (en) * 2000-04-17 2011-07-05 The University Of Sydney Method and apparatus for objective electrophysiological assessment of visual function
US8016416B1 (en) * 2005-01-15 2011-09-13 Sandy Helene Straus Automatic system and methods for measuring and evaluating at least one of mass vision, cognition, knowledge, operation skills, and the like
US8036896B2 (en) * 2006-04-18 2011-10-11 Nuance Communications, Inc. System, server and method for distributed literacy and language skill instruction
US8082152B2 (en) * 2005-03-31 2011-12-20 Erocca Device for communication for persons with speech and/or hearing handicap
US20120021390A1 (en) * 2009-01-31 2012-01-26 Enda Patrick Dodd Method and system for developing language and speech
US20120314045A1 (en) * 2009-08-26 2012-12-13 Ecole Polytechnique Federale De Lausanne (Epfl) Wearable systems for audio, visual and gaze monitoring
US20120329018A1 (en) * 2006-07-18 2012-12-27 Barry Katz Response scoring system for verbal behavior within a behavioral stream with a remote central processing system and associated handheld communicating devices
US20130021373A1 (en) * 2011-07-22 2013-01-24 Vaught Benjamin I Automatic Text Scrolling On A Head-Mounted Display
US20130031475A1 (en) * 2010-10-18 2013-01-31 Scene 53 Inc. Social network based virtual assembly places
US8392186B2 (en) * 2010-05-18 2013-03-05 K-Nfb Reading Technology, Inc. Audio synchronization for document narration with user-selected playback
US20130073387A1 (en) * 2011-09-15 2013-03-21 Stephan HEATH System and method for providing educational related social/geo/promo link promotional data sets for end user display of interactive ad links, promotions and sale of products, goods, and/or services integrated with 3d spatial geomapping, company and local information for selected worldwide locations and social networking
US8465153B1 (en) * 2010-11-23 2013-06-18 Joergen Bruun-Jensen System for clinical examination of visual functions using lenticular optics or programmable displays
US8475391B2 (en) * 2009-09-16 2013-07-02 Cerebral Assessment Systems Method and system for quantitative assessment of spatial distractor tasks
US8562541B2 (en) * 2009-09-16 2013-10-22 Cerebral Assessment Systems, Inc. Method and system for quantitative assessment of visual motion discrimination
US20140049462A1 (en) * 2012-08-20 2014-02-20 Google Inc. User interface element focus based on user's gaze
US8690325B1 (en) * 2005-07-12 2014-04-08 Sandy Helene Straus Sensory input devices, sensory output devices, and automatic systems, methods, and apparatuses for at least one of mass measurement, evaluation, or communication
US8714987B2 (en) * 2007-03-28 2014-05-06 Breakthrough Performancetech, Llc Systems and methods for computerized interactive training
US8740794B2 (en) * 2010-10-21 2014-06-03 Queens' University At Kingston Method and apparatus for assessing or detecting brain injury and neurological disorders
US8777630B2 (en) * 2009-09-16 2014-07-15 Cerebral Assessment Systems, Inc. Method and system for quantitative assessment of facial emotion sensitivity
US20140220520A1 (en) * 2011-09-09 2014-08-07 Articulate Technologies Inc. Intraoral tactile feedback methods, devices, and systems for speech and language training
US20140356822A1 (en) * 2013-06-03 2014-12-04 Massachusetts Institute Of Technology Methods and apparatus for conversation coach
US20150099946A1 (en) * 2013-10-09 2015-04-09 Nedim T. SAHIN Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device
US9072478B1 (en) * 2013-06-10 2015-07-07 AutismSees LLC System and method for improving presentation skills

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8048002B2 (en) * 2004-04-27 2011-11-01 Jamshid Ghajar Method for improving cognition and motor timing

Patent Citations (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6045515A (en) * 1997-04-07 2000-04-04 Lawton; Teri A. Methods and apparatus for diagnosing and remediating reading disorders
US5920838A (en) * 1997-06-02 1999-07-06 Carnegie Mellon University Reading and pronunciation tutor
US6931587B1 (en) * 1998-01-29 2005-08-16 Philip R. Krause Teleprompter device
US7211050B1 (en) * 1999-06-15 2007-05-01 Dimitri Caplygin System for enhancement of neurophysiological processes
US7149690B2 (en) * 1999-09-09 2006-12-12 Lucent Technologies Inc. Method and apparatus for interactive language instruction
US7972278B2 (en) * 2000-04-17 2011-07-05 The University Of Sydney Method and apparatus for objective electrophysiological assessment of visual function
US20010046659A1 (en) * 2000-05-16 2001-11-29 William Oster System for improving reading & speaking
US7254531B2 (en) * 2000-09-05 2007-08-07 Nir Einat H In-context analysis and automatic translation
US6704699B2 (en) * 2000-09-05 2004-03-09 Einat H. Nir Language acquisition aide
US20020099305A1 (en) * 2000-12-28 2002-07-25 Matsushita Electic Works, Ltd. Non-invasive brain function examination
US7678047B2 (en) * 2001-11-13 2010-03-16 Electronic Navigation Research Institute Chaologic brain function diagnosis apparatus
US20060093998A1 (en) * 2003-03-21 2006-05-04 Roel Vertegaal Method and apparatus for communication between humans and devices
US20050119894A1 (en) * 2003-10-20 2005-06-02 Cutler Ann R. System and process for feedback speech instruction
US20060270945A1 (en) * 2004-02-11 2006-11-30 Jamshid Ghajar Cognition and motor timing diagnosis using smooth eye pursuit analysis
US20050273017A1 (en) * 2004-03-26 2005-12-08 Evian Gordon Collective brain measurement system and method
US7815507B2 (en) * 2004-06-18 2010-10-19 Igt Game machine user interface using a non-contact eye motion recognition device
US7697825B2 (en) * 2004-08-18 2010-04-13 Sunplus Technology Co., Ltd. DVD player with language learning function
US8016416B1 (en) * 2005-01-15 2011-09-13 Sandy Helene Straus Automatic system and methods for measuring and evaluating at least one of mass vision, cognition, knowledge, operation skills, and the like
US8082152B2 (en) * 2005-03-31 2011-12-20 Erocca Device for communication for persons with speech and/or hearing handicap
US8690325B1 (en) * 2005-07-12 2014-04-08 Sandy Helene Straus Sensory input devices, sensory output devices, and automatic systems, methods, and apparatuses for at least one of mass measurement, evaluation, or communication
US20070050151A1 (en) * 2005-08-10 2007-03-01 Shinji Satoh Psychotic manifestation and mental state evaluation apparatus and evaluation method
US20070166676A1 (en) * 2005-12-15 2007-07-19 Posit Science Corporation Cognitive training using guided eye movements
US20070248938A1 (en) * 2006-01-27 2007-10-25 Rocketreader Pty Ltd Method for teaching reading using systematic and adaptive word recognition training and system for realizing this method.
US20070218432A1 (en) * 2006-03-15 2007-09-20 Glass Andrew B System and Method for Controlling the Presentation of Material and Operation of External Devices
US8036896B2 (en) * 2006-04-18 2011-10-11 Nuance Communications, Inc. System, server and method for distributed literacy and language skill instruction
US20120329018A1 (en) * 2006-07-18 2012-12-27 Barry Katz Response scoring system for verbal behavior within a behavioral stream with a remote central processing system and associated handheld communicating devices
US20090051877A1 (en) * 2006-12-15 2009-02-26 Posit Science Corporation Cognitive training using guided eye movements
US8714987B2 (en) * 2007-03-28 2014-05-06 Breakthrough Performancetech, Llc Systems and methods for computerized interactive training
US20080309889A1 (en) * 2007-06-18 2008-12-18 Keynote Products Llc Speech Prompter Apparatus
US20090089062A1 (en) * 2007-10-01 2009-04-02 Fang Lu Public speaking self-evaluation tool
US20100092929A1 (en) * 2008-10-14 2010-04-15 Ohio University Cognitive and Linguistic Assessment Using Eye Tracking
US20110026779A1 (en) * 2008-12-24 2011-02-03 David Matsumoto Systems and methods for analyzing facial expressions, identifying intent and transforming images through review of facial expressions
US20120021390A1 (en) * 2009-01-31 2012-01-26 Enda Patrick Dodd Method and system for developing language and speech
US20120314045A1 (en) * 2009-08-26 2012-12-13 Ecole Polytechnique Federale De Lausanne (Epfl) Wearable systems for audio, visual and gaze monitoring
US20110063571A1 (en) * 2009-09-16 2011-03-17 Duffy Charles J Method and system for quantitative assessment of visual contrast sensitivity
US20110065069A1 (en) * 2009-09-16 2011-03-17 Duffy Charles J Method and system for quantitative assessment of verbal recognition memory
US8777630B2 (en) * 2009-09-16 2014-07-15 Cerebral Assessment Systems, Inc. Method and system for quantitative assessment of facial emotion sensitivity
US8475391B2 (en) * 2009-09-16 2013-07-02 Cerebral Assessment Systems Method and system for quantitative assessment of spatial distractor tasks
US8562541B2 (en) * 2009-09-16 2013-10-22 Cerebral Assessment Systems, Inc. Method and system for quantitative assessment of visual motion discrimination
US20110123967A1 (en) * 2009-11-24 2011-05-26 Xerox Corporation Dialog system for comprehension evaluation
US8392186B2 (en) * 2010-05-18 2013-03-05 K-Nfb Reading Technology, Inc. Audio synchronization for document narration with user-selected playback
US20130031475A1 (en) * 2010-10-18 2013-01-31 Scene 53 Inc. Social network based virtual assembly places
US8740794B2 (en) * 2010-10-21 2014-06-03 Queens' University At Kingston Method and apparatus for assessing or detecting brain injury and neurological disorders
US8465153B1 (en) * 2010-11-23 2013-06-18 Joergen Bruun-Jensen System for clinical examination of visual functions using lenticular optics or programmable displays
US20130021373A1 (en) * 2011-07-22 2013-01-24 Vaught Benjamin I Automatic Text Scrolling On A Head-Mounted Display
US20140220520A1 (en) * 2011-09-09 2014-08-07 Articulate Technologies Inc. Intraoral tactile feedback methods, devices, and systems for speech and language training
US20130073387A1 (en) * 2011-09-15 2013-03-21 Stephan HEATH System and method for providing educational related social/geo/promo link promotional data sets for end user display of interactive ad links, promotions and sale of products, goods, and/or services integrated with 3d spatial geomapping, company and local information for selected worldwide locations and social networking
US20140049462A1 (en) * 2012-08-20 2014-02-20 Google Inc. User interface element focus based on user's gaze
US20140356822A1 (en) * 2013-06-03 2014-12-04 Massachusetts Institute Of Technology Methods and apparatus for conversation coach
US9072478B1 (en) * 2013-06-10 2015-07-07 AutismSees LLC System and method for improving presentation skills
US20150099946A1 (en) * 2013-10-09 2015-04-09 Nedim T. SAHIN Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10732784B2 (en) 2016-09-01 2020-08-04 University Of Massachusetts System and methods for cuing visual attention
EP4018647A4 (en) * 2019-08-23 2023-08-16 Your Speech Factory AB Electronic device and method for eye-contact training

Also Published As

Publication number Publication date
US9072478B1 (en) 2015-07-07

Similar Documents

Publication Publication Date Title
US9072478B1 (en) System and method for improving presentation skills
US10395545B2 (en) Analyzing speech delivery
US9665567B2 (en) Suggesting emoji characters based on current contextual emotional state of user
US11012486B2 (en) Personalized video playback
WO2017070496A1 (en) Automatic test personalization
US9336268B1 (en) Relativistic sentiment analyzer
US10230680B2 (en) Intelligently splitting text in messages posted on social media website to be more readable and understandable for user
US20180286099A1 (en) Sparse-data generative model for pseudo-puppet memory recast
US9700200B2 (en) Detecting visual impairment through normal use of a mobile device
US20170061989A1 (en) Conversational analytics
US20220141266A1 (en) System and method to improve video conferencing using presence metrics
US11756567B2 (en) Autocreation of conversational image representation
US20200233925A1 (en) Summarizing information from different sources based on personal learning styles
EP3185523B1 (en) System and method for providing interaction between a user and an embodied conversational agent
US11798675B2 (en) Generating and searching data structures that facilitate measurement-informed treatment recommendation
US10379709B2 (en) Electronically analyzing user activity on a graphical user interface
Sun et al. Response time of young children with complex communication needs following a communication opportunity
US10602976B2 (en) Personalized posture correction
US10872289B2 (en) Method and system for facilitating context based information
US20230061210A1 (en) Method and system of automated question generation for speech assistance
US20180253882A1 (en) Generating a personal avatar and morphing the avatar in time
US20230055421A1 (en) Caption customization and editing
TWM447567U (en) Language learning system using video clips for role playing
US20200387816A1 (en) User activity based cognitive question generation
Chassidim et al. SPOT&SPOT: A VIDEO VISUAL SCENE DISPLAY APPLICATION TO ENHANCE COMMUNICATION FOR INDIVIDUALS WITH ASD

Legal Events

Date Code Title Description
AS Assignment

Owner name: AUTISMSEES LLC, SOUTH CAROLINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:FEERST, DANIELLE A.;REEL/FRAME:036339/0597

Effective date: 20140623

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION