WO2016077681A1 - Système et procédé pour un étiquetage avec une voix et une icône - Google Patents

Système et procédé pour un étiquetage avec une voix et une icône Download PDF

Info

Publication number
WO2016077681A1
WO2016077681A1 PCT/US2015/060552 US2015060552W WO2016077681A1 WO 2016077681 A1 WO2016077681 A1 WO 2016077681A1 US 2015060552 W US2015060552 W US 2015060552W WO 2016077681 A1 WO2016077681 A1 WO 2016077681A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
tag
tags
data
icon
Prior art date
Application number
PCT/US2015/060552
Other languages
English (en)
Inventor
Ian Sherrard THOMPSON
Adeel ARIF
Nadine Elizabeth THOMPSON
William Thompson
Original Assignee
Koobecafe, Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koobecafe, Llc filed Critical Koobecafe, Llc
Publication of WO2016077681A1 publication Critical patent/WO2016077681A1/fr

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/54Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present disclosure relates to a system, method, and software application for tagging contextual content and, more specifically but not exclusively, to a user interface for tagging various subjects with voice tags and/or icon tags in an electronic device.
  • Picture or Icon tagging defines a technique that uses an image icon to identify a group of items or information in a computer.
  • a folder displaying a selected picture may be created on the user device, which comprises of the items stored in reference to that image.
  • voice tagging defines a technique that enables a user to record a voice tag for a particular item or subject and associate the voice tag with it. Such tags are then used to retrieve the group of items or subjects referenced with the particular tags by the user.
  • the systems and methods can recognize the sequence of spoken words and translate it into contextual text and interpret the textual words by use of natural-language processing techniques and associate a tag with the requested word or phrase as desired by the user.
  • Embodiments of the present invention make the process of organizing, browsing and retrieving of data interactive, dynamic and user friendly by tagging it with voice and/or icon/images. It provides a system, method, and a software application, which allows a user to tag content with an icon or images.
  • the software application enables a user to easily associate an icon/image tag with a desired content, and intelligently process the icon/image tag and use it to retrieve the desired content for later use.
  • the present invention provides a graphical user interface to the user, which enables him/her to tag and associate data with voice and/or icon/images.
  • Another object of the present invention is to redefine traditional methods for data storage and retrieval on conventional computer systems, electronic media and network storage systems.
  • the present invention provides a system, method and software application for tagging data/content which can be implemented on any electronic device, and particularly, a smartphone, mobile phone, digital photo-frame, computer, laptop, radio, television, car navigation system, digital cameras, any PDA, other electronic entertainment devices, wireless communication devices, etc.
  • a method of identifying data on a user device comprising the steps of: associating one or more data sets stored in a memory of the user device with one or more tags, each tag generated based on a respective user input, with the one or more tags being stored in a tag repository in the memory; matching a tag corresponding to a user request with a subset of the one or more tags stored in the tag repository; and retrieving from the memory at least one data set associated with any tag of the subset of the one or more tags.
  • a non-transitory machine-readable medium having encoded thereon program code, wherein, when the program code is executed by a machine, the machine implements a method of identifying data, the method comprising: associating one or more data sets stored in a memory wit one or more tags, each tag generated based on a respective user input, with the one or more tags being stored in a tag repository in the memory; matching a tag corresponding to a user request with a subset of the one or more tags scored in the tag repository; and retrieving from the memory at least one data set associated with any tag of the subset of the one or more tags.
  • FIG. 1 is a flowchart presenting the steps of generating a voice tag in accordance with an embodiment disclosed within this specification;
  • FIG. 2 is a flowchart presenting the steps of retrieving data/content associated with a voice tag in accordance with an embodiment disclosed within this specification;
  • FIG. 3 is block diagram representing the user interface used in icon/image tagging application in accordance with an embodiment disclosed within this specification;
  • FIG, 4 is a flowchart presenting the steps of generating an icon/image tag in accordance with an embodiment disclosed within this specification.
  • FIG. 5 is a flowchart presenting the steps of retrieving data/content associated with an icon/image tag in accordance with another embodiment disclosed within this specification.
  • Some embodiments of the present invention provide a software application to tag content with a user's voice.
  • the software application enables a user to easily associate a voice tag with a desired content, and intelligently process the voice tag and use it to retrieve the desired content for later use.
  • the content may be any form of digital data, such as one or more of: a document, a coupon, a photo, a music file, a calendar date, a contact stored in the phone, a location address, a map, a web address or a web page, a file folder, etc.
  • methods are provided for users to tag digital media with an audio recording of their voice by speaking a word or a phrase and also the ability for users to search and browse digital data files using these voice tags. It should be understood that a "user" is the person saying the word or phrase.
  • inventions of the present invention provide systems and methods for tagging photos, videos and other form of digital media with spoken audio.
  • the systems and methods disclosed herein include the ability to recognize the sequence of spoken words and translate it into contextual text and interpreting the textual words by use of natural-language processing techniques and associating a tag with the requested word or phrase as desired by the user. Subsequently, if the same or some other user speaks a closely matching sequence of the voice tag, the systems and method disclosed herein will again convert it into text format and use natural-language processing techniques to interpret the user's request and match it with the stored voice tags and can retrieve the digital media if a match exists. However, in one embodiment, the system has an option to disable this feature, so that only the original user can retrieve the content.
  • Embodiments of the present invention use natural-language processing techniques to recognize the user's request and perform the requested action.
  • the voice tags associated with the software application are dynamic in nature, and the user has the ability to sec any spoken words into a voice tag and can customize it as desired.
  • FIG.l outlines the process of using the software application to generate and associate data with voice tags, in one embodiment of the invention.
  • the user wishes to generate a voice tag for a given piece of data on his or her user device.
  • the user device is any electronic device with means of displaying content or data to the user.
  • the user device can be but is not limited to a smartpbone, mobile phone, digital photo-frame, computer, laptop, radio, car navigation system, digital camera, television, any PDA, other electronic entertainment devices, wireless communication devices, etc.
  • the data to be taeged can be but are not limited to any form of digital data, such as documents, coupons, photos, music files, calendar dates, contacts stored in the phone, location addresses, maps, web addresses or web pages, file folders, etc.
  • the user holds onto (e.g., long-presses) the selected field on the screen or a button, depending on the device, for a fixed time interval to enable the tagging process to start.
  • the fixed time interval can be customized per the user's requirements. In one example embodiment of a smartphone application, the fixed time interval is approximately 2 seconds.
  • the method of holding onto the selected field can be performed in a variety of ways,
  • the user can touch the screen of his device and hold the displayed field or can just click or point at the data for the fixed time interval.
  • Other ways of selecting the field to be tagged can be a tap on the screen of the user device for a fixed number of times or to press a button of the user device for a pre-defined time interval with the field being available to the user.
  • the user may also have a shortcut option displayed on the screen to select the field for associating it with a voice tag.
  • the method of activating voice tagging can be implemented with simple voice commands, where the system interprets the voice to be a command to activate tagging,
  • the user hears two distinctive beeps, alerting him or her that the system is ready for tagging.
  • the user can also customize the beeps to a digital voice command telling the user to be prepared to record the voice tag.
  • a voice can alert the user saying "ready to proceed with voice tagging.”
  • the user speaks an audio request comprising a system command to perform a certain action and a voice tag to associate it with the selected data.
  • the step 104 of speaking the audio request is initiated by the user after the two beeps have been played or a voice command has alerted the user.
  • the spoken audio request may include any audible word, phrase, number, and/or any sound capable of being recorded by the user device, Further, the user device incorporates means to receive the audio request spoken by the user by way of a microphone and/or any other desired input device capable of receiving the spoken audio of the user. As an option, the user audio may be received utilizing an input device including an integrated microphone in a set of headphones. As another option, the user audio may be received utilizing an input device including a wireless device which can be positioned for optimum voice control in an automobile, an indoor environment, and/or an o u tdoor environment.
  • the audio request spoken by the user is received by the software application and is further processed at step 105, wherein the user's audio is converted into contextual text,
  • the system interprets the audio request by recognizing a portion of the sequence of words (text format) as system commands, and the remaining portion as the requested voice tag by use of natural-language processing techniques.
  • the data is tagged with the user-requested word or phrase.
  • the voice tag may be verified by replaying the recorded voice for the user who provided the spoken audio and allowing the user to either accept or reject the voice tag. For example, in playing the voice tag back for the user, the user may be prompted to either state "yes" or "no.” If the user rejects the set voice tag, then the user may subsequently be prompted to record the voice tag again. In this way, the verifying step may allow for the adjusting of the received voice tag. Any other type of verification process may optionally be utilized for verifying the recorded voice tag.
  • a photo is displayed on the user device, and the user wishes to tag the photo
  • the user can touch the photo on the screen for two seconds followed by which he or she will hear two distinctive beeps, which signals the user that system is ready to tag the concent.
  • the user will speak an audio word or phrase, the received audio will be converted into text form, interpreted using the natural-language processing techniques to determine the user's request, and the photo will be tagged with the word/phrase as desired by the user.
  • the software application of the present invention may be installed on the user device or accessed over a network.
  • the system may employ any number of any conventional or other databases, data stores or storage structures (e.g., files, databases, data structures, data or other repositories, etc.) to store information (e.g. displayed data, voice tags).
  • the database system may be included within or coupled to the server and/or client systems or devices.
  • the database systems and/or storage structures may be remote from or local to the device.
  • Embodiments of the present invention may be utilized to tag any type of data object (e.g., still image, picture, video, multimedia object, audio, music file, etc.).
  • the voice tags may include any voice and/or speech signals containing any desired information pertaining to the data to be tagged.
  • the data may be associated with any quantity of voice tags.
  • the voice tags may include any desired information pertaining to any entity.
  • the entity may include any desired object (e.g., person, animal, animate or inanimate object, any item in a social network that can be associated with a voice tag, etc.).
  • Embodiments of the present invention may be employed with any suitable social media or other environment employing tagging of objects.
  • the present invention provides a system with the capability to interact with the user and understand the user's request by using natural-language processing techniques, which makes the software application dynamic in nature and easier to use.
  • the process of retrieving the data associated with a saved voice tag is described in detail below, for one embodiment of the invention:
  • the user wishes to retrieve stored data he or she has linked with a particular voice tag.
  • the user holds onto the screen of the user device for a fixed time interval.
  • the fixed time interval can be customized per the user's needs or system requirements.
  • the user can press any pre-defined button on the display screen of the user device according to a fixed algorithm or can use a shortcut key or option for this purpose.
  • the system responds with two distinctive beeps to the user action, alerting the user that the system is ready for the voice tag retrieval process.
  • the user can also customize the beeps to a digital voice command telling the user that voice tag retrieval process is being executed.
  • a voice can alert the user saying "'system ready to find voice tag.”
  • the number of pre-defined beeps could be different per the user's needs or system requirements.
  • the user speaks an audio request comprising a system command and the saved voice tag the user has earlier stored and associated with selected data.
  • the audio request consists only of the stored voice tag, with no system command.
  • the software application has the ability to recognize the audio request and take action according to it.
  • the software application receives the audio request as spoken by the user and converts it into text using a voice-to- text conversion application.
  • the software application interprets a portion of the sequence of words (text format) as system commands and the remaining portion as the voice tag stored in the repository by using natural-language processing techniques. For instance, if a user says, “Retrieve Vacation Photos,” then the system would interpret "Retrieve” as a system command and "'Vacation Photos” as a tag.
  • the system matches the voice tag recognized in the user's request with those stored in the repository of the system, and if there exists a match, then the data associated with the voice tag is retrieved and made available to the user as indicated at step 208.
  • the repository where the voice tags are stored may be located in the user device or on the network, where it is accessible by the user. If a match does not exist for the voice tag as spoken by the user, then the system prompts the user to repeat the above steps.
  • the user's audio request consists only of the stored voice tag, with no system command. In such a case, if the system cannot interpret any portion of the audio request as a system command, then the complete sequence of the audio request is considered to be the voice tag.
  • Some embodiments of the present invention broadly follow the same methodology for generating new voice tags, as well as for the retrieval of stored voice tag data, which makes the software application more user-friendly, desirable, and time-efficient.
  • voice tags are further accessible for data mining/text analytics. Moreover, voice tags are language-dependent, but managed In a language-oriented manner, and may be cross-linked in Enterprise Content Management (ECM) environments.
  • ECM Enterprise Content Management
  • the matching of voice tags may not be in the exact sequence as earlier stored by the user. More specifically, if the spoken words of the user match all the words of the stored voice tag in a random order, then it is still considered a match, and the files associated with the tag are displayed to the user.
  • the system uses advanced natural-language processing techniques and users' previous interactions with the application to learn and adapt to the users' choice and order of words.
  • the files associated with it are displayed to the user. For example, if the user has earlier tagged the photos of a person with a voice tag "'young John” and later executes the application to find a tag by saying "John” only, then it is still considered a match with the voice tag '"young John," and all the photos associated with the voice tag are displayed to the user.
  • the recorded voice taes can be processed and automatically converted into text for storage and presentation as a written transcript with the files when they are retrieved to be displayed to the user.
  • Embodiments of the present invention provide an innovative tagging methodology that makes it fun for users to tag media items such as pictures, videos, and contact information relating to people, times, events, and other relevant criteria.
  • the files may be automatically or manually shared on a social networking site or to a group within a social networking site based on the tags.
  • File shared on a social news feed can also be searchable using voice tag.
  • Voice tag-based search can function like a hashtag search, where a user can search using either te t or voice.
  • voice tags used can be popular branded one-liners from famous shows, artists and/or movies, or the like, allowing users to tag their content with popular branded one-liners. This can create entertainment value for the users, while allowing brands and content creators to use their one-liners as a potential marketing or monetization tool.
  • a system, method, and software application which allows a user to tag content with an icon or images.
  • the software application enables a user to easily associate an icon/image tag with a desired content, and intelligently process the icon/image tag and use it to retrieve the desired content for later use.
  • the content may be of any form of digital data, such as documents, coupons, photos, music files, calendar dates, contacts stored in the phone, location addresses, maps, web addresses or web pages, file folders, etc.
  • the method gives a user the option to tag content with images that relate to the specific content - for example, tagging travel-related content with a palm tree icon/image or shopping-related content with a basket icon/image etc.
  • methods are provided for users to tag digital media with an icon/image tag and the ability for users to search and browse digital data files using these icon/image tags.
  • the software application provides the user with a set of icons displayed on the user device that can be used for tagging purposes.
  • the user can also customize the icons available for tagging and has the ability to add personal icons as well.
  • a graphical user interface interacts with the user to enable tagging of content with the icon/image of his choice.
  • GUI views may be displayed in a GUI and may be configured to enable a user to interact with digital image tagging tools, features, and functionality.
  • the GUI may include an image tagging view and/or other tool(s) configured to enable a user to provide user input for tagging content with a particular icon/image.
  • a plurality of icons are listed on a dial arranged in a semi-circular design on the display of the user device that can be swiped in and out of the phone screen.
  • the semi-circular dial design allows the users to have a large number of icons accessible in a relatively small space, and users can browse through the plurality of icons displayed on the dial and select the preferred icon by turning or rotating the dial.
  • the method of displaying the plurality of icons for user selection can be implemented in a variety of ways.
  • a drop-down list of icons could be made available to the user, or an icon search box can be provided in which the user can just type words related to the icon he or she wishes to tag with the desired content.
  • the system will search in the repository or on the network for the icons that match the user choice and display it to the user. Any geometric or three-dimensional GUI can be used to display the plurality of icons for user selection to be used in the tagging process.
  • FIG. 3 shows the semi-circular dial 10, which is utilized to provide the user with a plurality of icons for tagging process, in one embodiment of the invention.
  • the dial 10 is displayed on the user device and includes five different icons 10a, 10b, 10c, lOd, and lOe, as indicated in FIG. 3.
  • the dial 10 may comprise many other icons and can be rotated clockwise or counterclockwise with the touch of the user on the screen panel to browse through other icons. The rotation of the dial 10 can also be performed by a press of the button or through an option displayed to the user.
  • FIG. 3 shows the semi-circular dial 10, which is utilized to provide the user with a plurality of icons for tagging process, in one embodiment of the invention.
  • the dial 10 is displayed on the user device and includes five different icons 10a, 10b, 10c, lOd, and lOe, as indicated in FIG. 3.
  • the dial 10 may comprise many other icons and can be rotated clockwise or counterclockwise with the touch of the user on the
  • buttons 10a, 10b, 10c, lOd, and lOe are displayed together on the dial 10, but the number of icons to be displayed in a single set can be pre-defined by the user according to his/her convenience.
  • the dial can also be zoomed or minimized by the user while using the software application.
  • Icon 10c may be displayed in a different color or alternatively marked, as compared to the other icons, 10a, 10b, lOd, and lOe, which indicates that this icon is currently selected by the user.
  • a text box could be displayed with the icon, indicating its current selection by the user.
  • the process of tagging content/data with a certain icon or image is described below in detail:
  • the user wishes to tag particular data with an image/icon.
  • the data to be tagged can be but is not limited to any form of digital data, such as documents, coupons, photos, music files, calendar dates, contacts stored in the phone, location addresses, maps, web addresses or web pages, file folders, etc.
  • the method of swiping out the dial by the user can include, e.g., the steps of touching the device screen and dragging the user's finger on the screen in a direction to enable the dial to be displayed.
  • the method of displaying the dial on the screen can be pre-defined and customized by the user.
  • the user device can be any electronic device with means of displaying a user interface with the content or data to the user.
  • the user device can be but is not limited to a smartphone, mobile phone, digital photo-frame, computer, laptop, radio, car navigation system, digital camera, television, any PDA, other electronic entertainment device, wireless communication device, etc.
  • the user rotates the dial to browse through the plurality of icons available on the dial and finally selects an icon the user wants to tag with the content or data.
  • the rotation of the dial can be clockwise and counterclockwise, and the user has the option to rotate the dial at a speed per his/her convenience.
  • the selected icon can be displayed in a different color, compared to the other icons, to indicate the user choice.
  • the user double-taps the selected icon to tag with the displayed data on the user screen.
  • the step 404 of double-tapping enables the software application to trigger the action of tagging the displayed data on the user screen with the user selected icon/image.
  • the dial is displayed on the user device, the data to be tagged is made available in the background. Further, multiple data files or entities can be tagged with a single icon by the user.
  • the data to be tagged can be selected earlier or later by the user after performing steps 401 to 404.
  • the system of the present invention processes the user's request of tagging the data with a particular icon and takes a screenshot of the displayed data, saves the URL/link of the data, and/or saves the content of the file displayed on the screen.
  • the system is capable of performing any one of these actions or a combination. By initiating one or a combination of these tasks, the user- selected icon/image is tagged with the displayed data on his or her user device.
  • the data corresponding to the icons tagged with the user-selected data can be stored on the user device in a repository or can be saved on the network.
  • the method of the present invention to tag selected data with icons/images provides the user a very fast, efficient, and interactive application to manage and organize his or her data. It is easier to associate data entities with icons resembling the same subject, and the user will be able to retrieve the data more easily in the future, rather than searching all files and folders to reach the data that the user wants.
  • system of the present invention allows the user to tag multiple data files/content with a certain icon/image.
  • the user is looking at a travel gallery folder on his smartphone that contains photos, videos, and other files relevant to his visit to Florida, and the user wants to tag using a travel icon, e.g., a palm tree.
  • a travel icon e.g., a palm tree.
  • the user will swipe out the dial on the phone screen while the folder is available in the background and will rotate the dial until he or she finds an image which he or she wants to tag with the travel folder (a palm tree in this example).
  • the user will double tap on the icon to tag it with the selected travel folder.
  • the system will process the user's request and will save the link location of the travel folder with the "palm tree" icon.
  • the user In case the user fails to find any icon on the dial that he or she wants to tag with certain data, the user has the option to browse through picture libraries or use the Internet to download new icons and add them into the dial.
  • FIG. 5 shows an example icon tag retrieval process to be used by the user for accessing data he/she has earlier tagged with a particular icon/image, in one embodiment.
  • the user wishes to retrieve data associated with a particular icon/image tag.
  • the user swipes out the dial on the user display screen.
  • the method of displaying the dial on the screen can be predefined and customized by the user.
  • the user rotates the dial to browse through the plurality of icons available on the dial and finally stops at the icon the user has earlier used to tag with certain data.
  • the selected icon will have a different color or size to make the user aware of his or her choice.
  • the user holds the icon down for a fixed time interval.
  • the method of holding onto the selected icon/image can be performed in a variety of ways.
  • the user can touch the screen of his device and hold the selected icon/image on the dial or can just the click or point at the icon/image for the fixed time interval.
  • Other ways of holding the tagged icon/image can be a tap on the screen of the user device for a fixed number of times, or to press a button of the user device for a pre-defined time interval with the icon being selected on the dial.
  • the user may also have a shortcut option displayed on screen to press the tagged icon on the dial for a fixed time interval.
  • the user action will trigger the software application to respond to it, and the data or content tagged with the icon held by the user will be displayed to the user.
  • the step of providing the retrieved data can include providing a screenshot of the data, a URL/link to the tagged data, and/or displaying content of the data file tagged on the screen of the user device.
  • the data address in the repository or the network could also be provided to the user where he or she can easily access it.
  • the system of the present invention enables icon/images to function like folders on a conventional computer system.
  • An image/icon is mapped to a specific folder, and an icon is displayed to the user, allowing users to associate an icon with a content item, as opposed to the folder structure. This makes the application dynamic, interactive, and user-friendly.
  • data representative of a tagged icon/image is published to a digital icon/image storage and access subsystem over a network.
  • the digital icon/image storage and access subsystem may be configured to score and provide a plurality of user devices with access to tagged digital icon/images on the dial and the data associated with them.
  • tags may be defined by various users, applied to a certain data, and used to create one or more conversation threads between the users.
  • Digital data/content can be stored and organized based on user-defined criteria, Dynamic organization of data/content based on interactive parameters can be provided.
  • Text-based descriptive data can be associated with nontext data.
  • Organization, storage, and retrieval of data/content by descriptive parameters can be provided.
  • Descriptive information can be associated with stored data without altering the data's contents. The system provides the user with more liberal decisionmaking, while accessing and storing data/content relevant to the user.
  • a method of identifying data on a user device comprising the steps of: (A) associating (e.g., 101-107, FIG. 1; 401 -405, FIG, 4) one or more data sets stored in a memory of the user device with one or more tags, each tag generated based on a respective user input, with the one or more tags being stored in a tag repository in the memory; (B) matching (e.g., 205-207, FIG. 2; 502-504, FIG. 5) a tag corresponding to a user request with a subset of the one or more tags stored in the tag repository; and (C) retrieving (e.g., 208, FIG. 2; 505, FIG. 5) from the memory at least one data set associated with any tag of the subset of the one or more tags.
  • A associating (e.g., 101-107, FIG. 1; 401 -405, FIG, 4) one or more data sets stored in a memory of the user device with one or more tags, each tag generated based on a
  • the at least one data set comprises two or more data sets.
  • the at least one data set includes: a first data set associated with a first tag of the subset of the one or more tags; and a second data set associated with a second tag of the subset of the one or more tags, the second tag being different from the first tag.
  • the step of associating comprises generating (e.g., 104-107, FIG. 1) a first tag for a first data set based on a first audio input.
  • the step of associating further comprises generating (e.g., 402-405, FIG. 4) a second tag for the first data set based on an image or icon.
  • the step of generating comprises: receiving (e.g., 104, FIG. 1) the first audio input; converting (e.g., 105, FIG. 1) the first audio input into a corresponding first text; and generating (106-107, FIG. 1) the first tag based on the corresponding first text.
  • the step of converting comprises: interpreting (e.g., 106, FIG. 1) the first audio input to identify a first portion and a second portion thereof, the first portion being a user command portion, and the second portion being a portion to be converted into the corresponding first text.
  • the step of matching comprises: receiving (e.g., 204, FIG. 2) a second audio input, the second audio input being the user request; converting (e.g., 205, FIG. 2) the second audio input into a corresponding second text; and identifying (e.g., 206-207, FIG. 2) the subset of the one or more tags by comparing the corresponding first text and the corresponding second text.
  • the step of associating comprises generating (e.g., 402-405, FIG. 4) a first tag for a first data set based on an image or icon.
  • the image or icon is selected based on the respective user input (e.g., 402-404, FIG. 4).
  • the step of generating comprises providing (e.g., as indicated in FIG. 3) a plurality of images or icons for user selection using a movable image dial (e.g., 10, FIG. 3) displayed on the user device.
  • a movable image dial e.g. 10, FIG. 3
  • the method further comprises providing a user interface for receiving user inputs.
  • the user interface comprises a graphical user interface (e.g., as indicated in FIG. 3).
  • the user device comprises at least one of: a smartphone, a digital photo-frame, a computer, a tablet, a digital book reader, a radio, a car navigation system, a digital camera, a television set, and a wireless communication device.
  • the one or more data sets include one or more of: a digital document, a digital coupon, a digital image file, a digital music file, a calendar date, an entry of a digital address book or contact list, a location address, a digital map, a web address, a web-page file or file set, and a file folder.
  • a non-transitory machine-readable medium having encoded thereon program code, wherein, when the program code is executed by a machine, the machine implements a method of identifying data, the method comprising the steps of: (A) associating (e.g., 101 -107, FIG. 1 ; 401-405, FIG. 4) one or more data sets stored in a memory with one or more tags, each tag generated based on a respective user input, with the one or more tags being stored in a tag repository in the memory; (B) matching (e.g., 205-207, FIG. 2; 502-504, FIG.
  • the hand-held or mobile user devices which can be used to implement the present invention are preferably equipped with a display or monitor, a base (e.g., including at least one processor, one or more memories and/or internal or external network interfaces or communications devices (e.g., wireless, etc.)), optional input devices (e.g., a keyboard, touch screen, or other input device), and any commercially available and custom software (e.g., communications software, capture module, browser/interface software, applications, etc.).
  • a base e.g., including at least one processor, one or more memories and/or internal or external network interfaces or communications devices (e.g., wireless, etc.)
  • optional input devices e.g., a keyboard, touch screen, or other input device
  • any commercially available and custom software e.g., communications software, capture module, browser/interface software, applications, etc.
  • the computer or other processing systems employed by the present invention embodiments may be implemented by any number of any personal or other type of computer or processing system (e.g., desktop, laptop, PDA, tablets or other mobile computing devices, etc.), and may include any commercially- available operating system and any combination of commercially available and custom software (e.g., browser software, communications software, server software, tag module, capture module, social media environment module, etc.).
  • the computer systems and devices may inc!ude any types of displays or monitors and input devices (e.g., keyboard, mouse, voice recognition, touch screen, etc.) to enter and/or view information.
  • the design of the dial as described in the present invention incorporates many advantages. Firstly, the semi -circular rotatable design profile of the dial allows large number of icons to be easily displayed and accessed by the user as opposed to conventional iPhone/smartphone devices, in which the display of the contents/icons on the user interface is limited by the width of the screen.
  • the dial design makes the user interface capable of providing a variety of icon/images displayed, while utilizing the device screen space efficiently.
  • the dial location can be set on the screen, per the user's comfort and needs.
  • the dial can be located on the right-side or left- side of the display panel for the right-handed or left-handed person's convenience and can also be situated on the top or the bottom side of the display screen.
  • the icons on the dial can be customized per the user's needs.
  • the user can create new icons and replace the existing icons with user generated new icons.
  • the user can also customize the icons into a folder, wherein one icon can hold a multiple set of icons.
  • a school icon on the dial can hold other multiple icons set in relation to it.
  • the dial switches to all the icons of the school folder, with the last icon beins the up button, and one icon being the home button.
  • the selection of the home button brings the user back to the main dial, whereas, when the up button is selected, the user returns to the icon dial hierarchy above the current icon dial.
  • processors may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software.
  • the functions may be provided by a single dedicated processor, by a single shared processor, or by a plurality of individual processors, some of which may be shared.
  • explicit use of the term "processor” or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (DSP) hardware, network processor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), read only memory (ROM) for storing software, random access memory (RAM ), and non volatile storage.
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • FPGA field programmable gate array
  • ROM read only memory
  • RAM random access memory
  • non volatile storage Other hardware, conventional and/or custom, may also be included.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

L'invention concerne un système, un procédé et une application logicielle pour étiqueter des données/contenus numériques avec une voix et/ou des icônes/images. L'utilisateur réalise une série d'actions sur son dispositif d'utilisateur pour permettre l'étiquetage de données/contenus avec une voix et/ou des icônes/images. Les étiquettes vocales sont des mots prononcés par un utilisateur associés à des entités de données spécifiées, tandis que les étiquettes d'icône/image sont des images ou icônes numériques sélectionnées par un utilisateur reliées à des ensembles de données stockés sur le dispositif d'utilisateur ou sur le réseau. Les étiquettes de voix et/ou d'icône sont définies d'une manière interactive et dynamique et traitées de manière intelligente par un système cohérent avec la présente invention et ultérieurement utilisées pour extraire un contenu relié à ces dernières.
PCT/US2015/060552 2014-11-14 2015-11-13 Système et procédé pour un étiquetage avec une voix et une icône WO2016077681A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201462079910P 2014-11-14 2014-11-14
US62/079,910 2014-11-14

Publications (1)

Publication Number Publication Date
WO2016077681A1 true WO2016077681A1 (fr) 2016-05-19

Family

ID=54609006

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2015/060552 WO2016077681A1 (fr) 2014-11-14 2015-11-13 Système et procédé pour un étiquetage avec une voix et une icône

Country Status (1)

Country Link
WO (1) WO2016077681A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108446339A (zh) * 2018-02-28 2018-08-24 维沃移动通信有限公司 一种应用图标的分类方法及移动终端
CN109145204A (zh) * 2018-07-27 2019-01-04 苏州思必驰信息科技有限公司 画像标签生成和使用方法及***
EP3674923A1 (fr) * 2018-12-24 2020-07-01 Mario Tykve Procédé de mémorisation et de reproduction, relatives aux objets, des images numériques

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050161510A1 (en) * 2003-12-19 2005-07-28 Arto Kiiskinen Image handling
US20070174326A1 (en) * 2006-01-24 2007-07-26 Microsoft Corporation Application of metadata to digital media
GB2445837A (en) * 2007-01-16 2008-07-23 Era Digital Media Co Ltd Mobile multimedia content distribution and access
WO2011018634A1 (fr) * 2009-08-14 2011-02-17 All In The Data Limited Repérage d’un contenu d’images fixes et animées avec des métadonnées
US20130250139A1 (en) * 2012-03-22 2013-09-26 Trung Tri Doan Method And System For Tagging And Organizing Images Generated By Mobile Communications Devices
US20130325462A1 (en) * 2012-05-31 2013-12-05 Yahoo! Inc. Automatic tag extraction from audio annotated photos
US20130346068A1 (en) * 2012-06-25 2013-12-26 Apple Inc. Voice-Based Image Tagging and Searching
US20140047386A1 (en) * 2012-08-13 2014-02-13 Digital Fridge Corporation Digital asset tagging
US20140081633A1 (en) * 2012-09-19 2014-03-20 Apple Inc. Voice-Based Media Searching

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050161510A1 (en) * 2003-12-19 2005-07-28 Arto Kiiskinen Image handling
US20070174326A1 (en) * 2006-01-24 2007-07-26 Microsoft Corporation Application of metadata to digital media
GB2445837A (en) * 2007-01-16 2008-07-23 Era Digital Media Co Ltd Mobile multimedia content distribution and access
WO2011018634A1 (fr) * 2009-08-14 2011-02-17 All In The Data Limited Repérage d’un contenu d’images fixes et animées avec des métadonnées
US20130250139A1 (en) * 2012-03-22 2013-09-26 Trung Tri Doan Method And System For Tagging And Organizing Images Generated By Mobile Communications Devices
US20130325462A1 (en) * 2012-05-31 2013-12-05 Yahoo! Inc. Automatic tag extraction from audio annotated photos
US20130346068A1 (en) * 2012-06-25 2013-12-26 Apple Inc. Voice-Based Image Tagging and Searching
US20140047386A1 (en) * 2012-08-13 2014-02-13 Digital Fridge Corporation Digital asset tagging
US20140081633A1 (en) * 2012-09-19 2014-03-20 Apple Inc. Voice-Based Media Searching

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108446339A (zh) * 2018-02-28 2018-08-24 维沃移动通信有限公司 一种应用图标的分类方法及移动终端
CN109145204A (zh) * 2018-07-27 2019-01-04 苏州思必驰信息科技有限公司 画像标签生成和使用方法及***
CN109145204B (zh) * 2018-07-27 2021-01-15 苏州思必驰信息科技有限公司 画像标签生成和使用方法及***
EP3674923A1 (fr) * 2018-12-24 2020-07-01 Mario Tykve Procédé de mémorisation et de reproduction, relatives aux objets, des images numériques

Similar Documents

Publication Publication Date Title
US11157577B2 (en) Method for searching and device thereof
US10521189B1 (en) Voice assistant with user data context
US10152199B2 (en) Object based contextual menu controls
US9190052B2 (en) Systems and methods for providing information discovery and retrieval
CN108369600B (zh) web浏览器扩展
US20140115070A1 (en) Apparatus and associated methods
US20160283055A1 (en) Customized contextual user interface information displays
US20090327272A1 (en) Method and System for Searching Multiple Data Types
KR20220000953A (ko) 터치스크린 상에 표시되는 조치 가능한 콘텐츠
JP2018504727A (ja) 参考文書の推薦方法及び装置
US11423113B2 (en) Contextual deep bookmarking
JP2017517052A (ja) 表示内容に対応する検索結果の提供
KR20160015838A (ko) 컨텐츠 분류 방법 및 이를 위한 디바이스
WO2016091095A1 (fr) Procédé et système de recherche basés sur une opération de toucher sur une interface de terminal
US20160179899A1 (en) Method of providing content and electronic apparatus performing the method
US11314826B2 (en) Method for searching and device thereof
US20140372467A1 (en) Contextual smart tags for content retrieval
US11586690B2 (en) Client-side personalization of search results
CN107015979B (zh) 一种数据处理方法、装置和智能终端
WO2019057191A1 (fr) Procédé de récupération de contenu, terminal et serveur, dispositif électronique et support d'informations
US20150067581A1 (en) System and method for a graphical user interface for hierarchical navigation
CN106558311B (zh) 语音内容提示方法和装置
KR20130080713A (ko) 음성 인식 기능을 구비한 이동 단말기 및 그 검색 결과 제공 방법
WO2016077681A1 (fr) Système et procédé pour un étiquetage avec une voix et une icône
US20180089150A1 (en) Providing location-based font recommendations

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15797805

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15797805

Country of ref document: EP

Kind code of ref document: A1