CN113590864A - Method and device for obtaining search result, electronic equipment and storage medium - Google Patents

Method and device for obtaining search result, electronic equipment and storage medium Download PDF

Info

Publication number
CN113590864A
CN113590864A CN202010367988.0A CN202010367988A CN113590864A CN 113590864 A CN113590864 A CN 113590864A CN 202010367988 A CN202010367988 A CN 202010367988A CN 113590864 A CN113590864 A CN 113590864A
Authority
CN
China
Prior art keywords
target object
target
image
topic
search result
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010367988.0A
Other languages
Chinese (zh)
Inventor
袁项南
陈轶博
李明伟
祖华龙
贾东雯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Baidu Online Network Technology Beijing Co Ltd
Shanghai Xiaodu Technology Co Ltd
Original Assignee
Baidu Online Network Technology Beijing Co Ltd
Shanghai Xiaodu Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Baidu Online Network Technology Beijing Co Ltd, Shanghai Xiaodu Technology Co Ltd filed Critical Baidu Online Network Technology Beijing Co Ltd
Priority to CN202010367988.0A priority Critical patent/CN113590864A/en
Publication of CN113590864A publication Critical patent/CN113590864A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Business, Economics & Management (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • User Interface Of Digital Computer (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The application discloses a method and a device for obtaining a search result, electronic equipment and a storage medium, and relates to the technical field of voice. The specific implementation scheme is as follows: receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of a target object in a preset area; acquiring an image of a preset area; and acquiring a search result of the target object according to the image. According to the method and the device, the user can enable the terminal device to automatically shoot the image through the voice command, and the search result of the target object is obtained, so that the method and the device are simple to operate and high in applicability.

Description

Method and device for obtaining search result, electronic equipment and storage medium
Technical Field
The present application relates to the field of image processing technologies, and in particular, to a method and an apparatus for obtaining a search result in a speech technology, an electronic device, and a storage medium.
Background
Students often encounter difficult problems in homework making process without explaining, and parents or teachers need to check the mistakes of homework after homework making, which takes a lot of time. To solve these problems, researchers have developed applications with which users can search for answers to questions and also check for errors in jobs.
In the prior art, a user needs to download an application program in a terminal device, shoot an image of a topic to be searched in the application program, the terminal device sends the image of the topic to a background server, the server obtains an answer of the topic from a topic library according to the topic in the image and feeds the answer back to the terminal device, and then the terminal device displays the answer of the topic.
However, in the problem searching mode in the prior art, a user needs to manually shoot the image of the problem in the application program, so that the operation is complex and the applicability is low.
Disclosure of Invention
The application provides a method and a device for obtaining a search result, electronic equipment and a storage medium, a user can enable the terminal equipment to automatically shoot an image through a voice instruction, the search result of a target object is obtained, and the method and the device are simple to operate and high in applicability.
A first aspect of the present application provides a method for obtaining search results, including: receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of a target object in a preset area; acquiring the preset area; and acquiring a search result of the target object according to the image.
In the method for obtaining the search result in the embodiment, the user can enable the terminal device to automatically obtain the image of the preset area through the voice instruction, and then obtain the search result of the target object according to the image.
A second aspect of the present application provides a method for obtaining search results, including: receiving an image from a terminal device, wherein the image is an image of a preset area, and the preset area comprises a target object; obtaining a search result of the target object according to the image; and sending the search result of the target object to the terminal equipment.
A third aspect of the present application provides a method for obtaining search results, including: receiving a target object from a terminal device; searching a database for a search result of the target object according to the target object; and sending the search result of the target object to the terminal equipment.
A fourth aspect of the present application provides an apparatus for obtaining a search result, including:
and the receiving and sending module is used for receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of the target object in the preset area.
And the processing module is used for acquiring the image of the preset area and acquiring the search result of the target object according to the image.
A fifth aspect of the present application provides an apparatus for obtaining search results, comprising:
and the transceiver module is used for receiving an image from the terminal equipment, wherein the image comprises a target object.
And the processing module is used for acquiring the search result of the target object according to the image.
The transceiver module is further configured to send the search result of the target object to the terminal device. Or,
the transceiver module in this embodiment is configured to receive a target object from a terminal device.
And the processing module is used for searching the search result of the target object in a database according to the target object.
The transceiver module is further configured to send the search result of the target object to the terminal device.
A sixth aspect of the present application provides an electronic apparatus comprising: at least one processor and memory; the memory stores computer-executable instructions; the at least one processor executes the computer-executable instructions stored by the memory to cause the electronic device to perform the methods of the first, second, and third aspects described above.
A seventh aspect of the present application provides a computer-readable storage medium having stored thereon computer-executable instructions that, when executed by a processor, implement the methods of the first, second and third aspects described above.
The application provides a method and a device for obtaining a search result, electronic equipment and a storage medium, and relates to the technical field of voice. The specific implementation scheme is as follows: the voice instruction indicates to acquire a search result of a target object in a preset area; acquiring an image of a preset area; and acquiring a search result of the target object according to the image. According to the method and the device, the user can enable the terminal device to automatically shoot the image through the voice command, and the search result of the target object is obtained, so that the method and the device are simple to operate and high in applicability.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The drawings are included to provide a better understanding of the present solution and are not intended to limit the present application. Wherein:
fig. 1 is a schematic view of a scene to which the method for obtaining a search result provided in the present application is applicable;
fig. 2 is a schematic view of a scenario in which a user uses a terminal device according to the present application;
FIG. 3 is a flowchart illustrating an embodiment of a method for obtaining search results provided by the present application;
fig. 4 is a first schematic diagram of an interface change of a terminal device provided in the present application;
FIG. 5 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present disclosure;
FIG. 6 is a first schematic diagram of an image including a target object according to the present disclosure;
FIG. 7 is a second schematic diagram of an image including a target object according to the present disclosure;
FIG. 8 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present disclosure;
FIG. 9 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present application;
FIG. 10 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present application;
fig. 11 is a second schematic diagram of an interface change of the terminal device provided in the present application;
FIG. 12 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present application;
FIG. 13 is a first schematic structural diagram of an apparatus for obtaining search results provided in the present application;
FIG. 14 is a second schematic structural diagram of an apparatus for obtaining search results provided in the present application;
fig. 15 is a schematic structural diagram of an electronic device provided in the present application.
Detailed Description
The following description of the exemplary embodiments of the present application, taken in conjunction with the accompanying drawings, includes various details of the embodiments of the application for the understanding of the same, which are to be considered exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present application. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
The intelligent equipment is various in types, such as an intelligent sound box, an intelligent robot, a learning machine and the like. The user realizes interaction with the intelligent equipment through a voice mode to obtain the search content required by the user, for example, the intelligent sound box can play songs, stories and the like according to the voice instruction of the user. Along with the development of intelligent equipment, increased a lot of interactive function in the intelligent equipment, if set up the display screen on intelligent audio amplifier, intelligent audio amplifier not only can play song, story, can also play video etc. according to user's instruction. In addition, if still be provided with the camera on the intelligent audio amplifier, can gather user's action, realize interacting with user's action, gather user's actions such as nodding the head, shaking the head, intelligent audio amplifier carries out corresponding operation according to user's action.
In the prior art, when searching for answers to questions, a user needs to open an application program in a terminal device, and then shoots images of the questions in the application program, the terminal device sends the images of the questions to a background server, the server obtains answers to the questions from a question bank according to the questions in the images and feeds the answers to the terminal device, and then the terminal device displays the answers to the questions. However, the problem searching method needs the user to manually shoot the image of the problem in the application program, is complex to operate, cannot effectively acquire the answer of the problem for the user who cannot operate the application program, and is low in applicability.
In order to solve the technical problem, the present application provides a method for obtaining a search result applied to a terminal device, so that the terminal device can automatically acquire an image of a target object through voice interaction between a user and the terminal device, and further obtain a search result of the target object, so as to display the search result of the target object. The method avoids the problem that a user needs to shoot the image of the target object manually, can acquire the image of the target object through voice, and is simple to operate and high in applicability.
Fig. 1 is a schematic view of a scene to which the method for obtaining a search result provided by the present application is applicable. As shown in fig. 1, a scenario to which the method for obtaining a search result provided by the present application is applicable includes: terminal equipment and server.
The terminal device in the present application may be, but is not limited to, a mobile terminal or a fixed terminal. The mobile terminal devices include, but are not limited to, smart speakers, learning machines, mobile phones, Personal Digital Assistants (PDAs), tablet computers, portable devices (e.g., laptop computers, pocket computers, or handheld computers), and the like. Fixed terminals include, but are not limited to, desktop computers and the like. In the method, the terminal equipment can perform voice interaction with the user, and the camera device and the display screen are integrated in the terminal equipment, so that when the user needs to acquire the search result of the target object, the image of the target object is shot through the camera device, and then the search result of the target object can be displayed on the display screen when the search result of the target object is acquired. Optionally, the terminal device may interact with the server to obtain the search result of the target object, and a specific interaction process is described in the following embodiments.
It should be understood that the target object in the present application may be a title, a text, a picture, or the like. For example, the method for obtaining the search result in the present application is applicable to a scene in which the user needs an answer to a search topic, a scene in which a search job is wrong, paraphrasing of a search word, or related information of an object in a search picture, and the like. The following embodiments are exemplified in the context of a user searching for an answer to a topic.
Fig. 2 is a schematic view of a scenario in which a user uses a terminal device according to the present application. Fig. 2 shows that when a user performs an operation, a book or a test paper is placed on a desktop, and a terminal device is placed on the desktop where the book or the test paper is far away from the user. It should be understood that the camera of the terminal device may be a camera rotating 360 degrees, and can acquire images of books or test papers; or, a camera array may be arranged in the terminal device to be able to acquire images in any direction or view angle, and further to acquire images of books or test papers; or, a camera (for example, on the top of the terminal device) may be arranged in the terminal device, and may not be capable of shooting images of books or test papers placed on the desktop. In the above, the setting manner of the camera in the terminal device is not limited in the present application, and the above example is only a realizable manner as long as the terminal device can acquire the image of the target object.
The following describes a method for obtaining search results provided by the present application with reference to specific embodiments. Fig. 3 is a flowchart illustrating an embodiment of a method for obtaining a search result according to the present application. The execution subject of the method flow shown in fig. 3 may be the terminal device or a chip, a processor, etc. in the terminal device. As shown in fig. 3, the method for obtaining search results provided by this embodiment may include:
s301, receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of a target object in a preset area.
S302, acquiring an image of a preset area.
S303, obtaining a search result of the target object according to the image.
In S301, the user may perform voice interaction with the terminal device, and after receiving the voice instruction of the user, the terminal device may analyze the semantic meaning of the voice instruction of the user, so as to execute a corresponding operation according to the semantic meaning of the voice instruction of the user. It should be understood that, in this embodiment, a method for how the terminal device parses the voice instruction of the user is not described in detail, and specifically, a voice interaction technology in the prior art may be referred to.
The voice instruction can instruct the terminal device to acquire a search result of the target object in the preset area. The preset area in this embodiment may be a shooting visual field range of the terminal device, that is, the voice instruction instructs the terminal device to obtain a search result of the target object by obtaining an image of the preset area. The target object in this embodiment may be a title, a text, a picture, or the like.
For example, the voice instruction may be "how to do the question", "what the word means", or "check my job", or "how to do the third question", and the voice instruction all requires the terminal device to obtain the search result of the target object according to the image of the preset area.
For example, when the voice instruction is "how to do this topic", the target object may be "this topic", the voice instruction is "check my job", and the target object may be "all topics in a book or a test paper currently acquired by the terminal device".
In the above S302, when the terminal device receives the voice instruction, the image of the preset area may be obtained, specifically, the terminal device may open the camera and control the camera to take a picture, so as to obtain the image of the preset area. It should be understood that the target object is included in the image of the preset area. It should be understood that, in the embodiment, at the position where the terminal device is placed in advance, the target object may be included in the shooting view of the camera of the terminal device.
In S303, in this embodiment, after the image is acquired, a search result of the target object may be acquired according to the image.
In this embodiment, optionally, after the terminal device captures an image of the preset region, the terminal device may display the image of the preset region, so that the user may see the image on the terminal device, and the terminal device may further display an operation frame, and the user may select the target object by enlarging or reducing the operation frame.
Fig. 4 is a first schematic view of an interface change of the terminal device provided in the present application. As shown in fig. 4, an image containing 3 titles is displayed on the interface 401 of the terminal device, and an operation box (shown by a dotted line box in the interface 401) is displayed, by which the user can select a 2 nd title, as shown by the interface 402, and accordingly, the 2 nd title is a target object.
In this embodiment, the terminal device may obtain a search result of the target object by using the target object. In a possible implementation manner, in this embodiment, the terminal device may obtain a search result of the target object in a local database. Illustratively, the terminal device identifies a target topic in the image, and obtains an answer to the target topic from a local database, where the answer to the target topic is a search result of the target object.
Or, in this embodiment, after the terminal device acquires the target object in the image, the terminal device may send the target object to the server, so that the server acquires the search result of the target object in the database of the server. Illustratively, after identifying a target topic in an image, a terminal device sends the target topic to a server, so that the server searches an answer of the target topic in a database according to the target topic, and the server can send the answer of the target topic to the terminal device, so that the terminal device obtains the answer of the target topic.
In a possible implementation manner, in this embodiment, the terminal device may also send the image of the target object to the server, so that the server identifies the target object in the image according to the image of the target object, for example, identifies a target topic in the image, so as to search for an answer to the target topic in the database, and the server may send the answer to the target topic to the terminal device, so that the terminal device obtains the answer to the target topic.
In this embodiment, after the terminal device obtains the search result of the target object, the search result of the target object may be displayed, so that the user may see the search result of the target object on the terminal device.
The method for obtaining the search result provided in the embodiment includes: receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of a target object; shooting an image containing a target object; and acquiring a search result of the target object according to the image. According to the method, the user can automatically shoot the image of the terminal device through the voice instruction and obtain the search result of the target object, the problem that the user needs to manually shoot the image of the target object is avoided, the image of the target object can be obtained through voice, the operation is simple, and the applicability is high.
On the basis of the above embodiment, a method for acquiring a target object in an image by a terminal device in the present embodiment is described below. Fig. 5 is a flowchart illustrating another embodiment of a method for obtaining search results provided by the present application. As shown in fig. 5, the method for obtaining search results provided by this embodiment may include:
s501, receiving a voice instruction of a user.
S502, acquiring an image of a preset area.
S503, acquiring the position of the target object indicated by the user in the image.
And S504, taking the object at the position as a target object.
And S505, acquiring a search result of the target object according to the target object.
The implementation in S501-S502 in this embodiment may refer to the related descriptions in S301-S302 in the above embodiments, which are not described herein again.
In S503 above, the terminal device may acquire the position of the target object indicated by the user in the image. In this embodiment, the position of the target object in the image, which is obtained by the terminal device and indicated by the user, is described from the following two possible scenarios:
the first scenario is: when the user inputs the voice instruction to the terminal device, the user can point at the target object with a finger to indicate the position of the target object in the image. Accordingly, the user's finger is included in the captured image.
The second scenario is: the voice instruction indicates a position of the target object in the image. For example, the voice instruction may be "how to do the top topic," where the top indicates the position of the target object in the image; alternatively, the voice instruction may be "what is the animal in the lower left corner", and the "lower left corner" indicates the position of the target object in the image.
In the above-described S504, in the present embodiment, since the user indicates the position of the target object in the image or in the voice instruction, the object at the position indicated by the user in the image can be taken as the target object. For example, a title at a position indicated by a user may be used as a target object, or text or a picture at the position indicated by the user may be used as the target object.
The target object in this embodiment may be a target title, and the title at the position indicated by the user is taken as the target object in this embodiment. However, although the user indicates the position of the target object, it is still necessary to acquire the boundary or coverage of the title at the position in the image to obtain the target title.
In this embodiment, the terminal device may identify the coverage of each topic in the image according to the identifier of each topic in the image, and then use the topic corresponding to the coverage to which the position belongs as the target topic. For example, in the first scenario, a title corresponding to a coverage to which the position of the fingertip of the user belongs may be used as the target title, or in the second scenario, a title corresponding to a coverage to which the position indicated in the user instruction belongs may be used as the target title.
Taking the first scenario as an example, fig. 6 is a first schematic diagram of an image including a target object provided in the present application. As shown in FIG. 6, the image contains three titles, and the positions of the user's fingertips are shown in the figure. In this embodiment, the terminal device may identify a coverage range of each topic in the image, which is shown by a dotted line frame in the image as the coverage range of each topic, and then the terminal device may use the topic 2 corresponding to the coverage range to which the position of the fingertip of the user belongs as the target topic.
In this scenario, in order not to obscure the target title, the user may place a finger at a blank position within the coverage of title 2, such as the illustrated a position.
Taking the second scenario as an example, fig. 7 is a schematic diagram of a second image containing a target object provided in the present application. As shown in fig. 7, the image includes three topics, and if the voice command of the user is "how to do the top topic", the topic on the mouth of the three topics can be determined to be the 1 st topic according to the arrangement of the topics in the diagram, and the 1 st topic is taken as the target topic.
In the above step S505, a search result of the target object is obtained according to the target object. In this embodiment, reference may be specifically made to fig. 8 and 9 for a manner in which the terminal device obtains a search result of the target object according to the target object. Fig. 8 is a flowchart illustrating another embodiment of a method for obtaining search results provided by the present application. Fig. 9 is a flowchart illustrating another embodiment of a method for obtaining search results provided by the present application.
As shown in fig. 8, S505 in the present embodiment may be replaced with: and acquiring a search result of the target object in the local database according to the target object. The target object in this embodiment is a target topic, and correspondingly, the search result of the target object is at least one of the following items: the answer of the target topic, the analysis of the target topic and the check result of the answer of the target topic. It should be understood that, in the example in this embodiment, the search result of the target object is taken as an example of the answer of the target topic.
As shown in fig. 9, in this embodiment, the terminal device may interact with the server, so that the server obtains the search result of the target object in the database of the server, and further feeds back the search result of the target object to the terminal device. Accordingly, S505 in the present embodiment may be replaced with S505 ', S505 ", and S505 '", where S505 ' is: the terminal device sends the target object to the server, and S505 ″ is: the server searches the database for the search result of the target object, and S505' ″ is: and sending the search result of the target object to the terminal equipment.
In this embodiment, the user may indicate the position of the target object in the image by means of finger indication or voice instruction indication, and then the terminal device may obtain the target object according to the position of the target object in the image indicated by the user, and further may obtain the search result of the target object.
On the basis of the above embodiment, another method for the terminal device to acquire the target object in the image in the present embodiment is described below. Fig. 10 is a flowchart illustrating a method for obtaining search results according to another embodiment of the present application. As shown in fig. 10, the method for obtaining search results provided by this embodiment may include:
s1001, receiving a voice command of a user.
S1002, acquiring an image of a preset area.
And S1003, acquiring the target object in the image according to the voice command.
S1004, according to the target object, obtaining a search result of the target object.
The implementation manners in S1001 to S1002 and S1004 in this embodiment may refer to the related descriptions in S301 to S302 and S505 in the foregoing embodiment, and are not described herein again.
In the above 1003, the voice command in the present embodiment indicates a target object. It should be understood that the target object in this embodiment is a target topic, and the voice command includes an identifier of the target topic or a text in the target topic to indicate the target object.
In this embodiment, the terminal device may recognize the identifier of the target topic or the text in the target topic in the voice command, and further obtain the target topic in the image according to the identifier of the target topic or the text in the target topic.
When the voice command includes the identification of the target topic, the identification of each topic in the image can be identified in this embodiment, and the topic identical to the identification of the target topic is taken as the target topic.
Illustratively, if the voice command is "how to do topic 2", the identification of the target topic in the voice command is 2, that is, the target object is indicated as topic 2. After recognizing the identification of the target topic in the voice command, the terminal device may recognize the identification of each topic in the image, for example, recognize the identifications "one", "two", and "three" of each topic in fig. 7, and further regard the 2 nd topic as the target topic.
When the voice command comprises characters in the target title, the terminal equipment can identify the characters in the image, and then takes the title, containing the characters in the target title, in the image as the target title to be selected, so as to determine the target title according to the target title to be selected.
It should be noted that the accuracy of determining the target topic by using the text in the target topic is relatively low, for example, there may be a plurality of topics including the text in the target topic. And if the target topic to be selected is one, taking the target topic to be selected as the target topic.
Illustratively, the voice command of the user is "to form a word with the following shape, and then the" to form a word with the following shape "is the text in the target topic, and accordingly, as shown in fig. 7, the target topic is the 2 nd topic.
In this embodiment, if there are multiple target topics to be selected, the target topics can be determined by interacting with the user. Specifically, in this embodiment, a plurality of target titles to be selected may be displayed, and the target titles may be determined according to a selection operation of a user on the plurality of target titles to be selected.
Illustratively, the user's voice command is "word group", and the "word group" is the text in the target topic. Correspondingly, as shown in fig. 7, it can be determined that the questions 2 and 3 both include the text "word group", and the target questions to be selected are the questions 2 and 3. In order to determine the final target topic, a plurality of target topics to be selected may be displayed in this embodiment. Fig. 11 is a schematic view illustrating an interface change of the terminal device according to the present application. As shown in an interface 1101 in fig. 11, the terminal device may display the 2 nd question and the 3 rd question, and the interface 1101 may also display a text reminding message of "a selection target topic" to remind the user to select the target topic. If the user clicks the 2 nd question, the interface 1101 jumps to the interface 1102, and the 2 nd question, the text reminding information of whether to determine the 2 nd question as the target question, and the "confirm" and "cancel" controls are displayed on the interface 1102, and if the user clicks the "confirm" control, the target question is the 2 nd question. It should be noted that the terminal device may also directly determine the 2 nd topic as the target topic without displaying the interface 1102. It should be noted that the interface 1101 is correspondingly displayed with selection controls behind the 2 nd question and the 3 rd question, so that the user clicks the selection controls to determine the target topic.
It should be understood that the manner of recognizing the text in the image and recognizing the identification of the title in the image in the present embodiment can refer to the prior art. It should be noted that in this embodiment, according to the manner of determining the coverage of the topic in the above embodiment, after identifying the identifier of the target topic or the text contained in the target topic, the target topic may be determined.
In this embodiment, the user may indicate the target topic in the voice instruction, so that the terminal device may determine the target topic in the image according to the voice instruction, and further obtain a search result of the target topic.
Fig. 12 is a flowchart illustrating another embodiment of a method for obtaining search results provided in the present application. As shown in fig. 12, the method for obtaining search results provided by this embodiment may include:
s1201, the terminal equipment receives a voice instruction of a user.
S1202, the terminal equipment acquires an image of a preset area.
S1203, the terminal device sends the image to a server.
S1204, the server obtains the search result of the target object according to the image.
And S1205, the server sends the search result of the target object to the terminal equipment.
The implementation in S1201-S1202 in this embodiment may refer to the description related to S301-S302 in the above embodiment, which is not described herein again.
Different from the foregoing embodiments, in all the foregoing embodiments, the terminal device determines the target object in the image after acquiring the image containing the target object. In this embodiment, after acquiring the image including the target object, the terminal device may send the image to the server, so that the server acquires the target object in the image. It should be understood that the image in the present embodiment includes the finger of the user, and the above S1204, that is, the manner in which the server acquires the target object in the image, and acquires the search result of the target object, may refer to the relevant description in the above embodiments.
In this embodiment, after acquiring an image including a target object, the terminal device may send the image to the server, and the server may determine the target object in the image, and further acquire a search result of the target object, and further feed back the search result of the target object to the terminal device.
Fig. 13 is a first schematic structural diagram of an apparatus for obtaining search results provided in the present application. The apparatus for obtaining the search result in this embodiment may be the terminal device in the foregoing embodiment. As shown in fig. 13, the apparatus 1300 for obtaining search results includes: a transceiver module 1301 and a processing module 1302.
The transceiving module 1301 is configured to receive a voice instruction of a user, where the voice instruction instructs to obtain a search result of a target object in a preset area.
The processing module 1302 is configured to obtain an image of the preset area, and obtain a search result of the target object according to the image.
In one possible design, the processing module 1302 is specifically configured to acquire a target object in an image; and acquiring a search result of the target object according to the target object.
In one possible design, the processing module 1302 is specifically configured to obtain a position of a target object indicated by a user in an image; the object at the position is taken as the target object.
In one possible design, the image includes a user's finger.
A processing module 1302, configured to obtain a position of a fingertip of the user in the image; the position of the user's fingertip is taken as the position of the target object in the image.
In one possible design, the voice instruction indicates a location of the target object in the image.
The processing module 1302 is specifically configured to determine a position of the target object in the image according to the voice instruction.
In one possible design, the target object is a target topic, and at least one topic is included in the image.
The processing module 1302 is specifically configured to identify a coverage area of each topic in the image according to an identifier of each topic in the image; and taking the title corresponding to the coverage range to which the position belongs as a target title.
In one possible design, the voice instruction indicates the target object.
The processing module 1302 is specifically configured to obtain a target object in an image according to a voice instruction.
In one possible design, the target object is a target topic, and the voice command includes an identifier of the target topic or a text in the target topic.
The processing module 1302 is specifically configured to identify an identifier of a target topic in the voice instruction or a character in the target topic; and acquiring the target topic in the image according to the identification of the target topic or the characters in the target topic.
In one possible design, at least one topic is included in the image.
A processing module 1302, configured to specifically identify an identifier of each topic in the image; and taking the topic with the same identification as the target topic.
In a possible design, the processing module 1302 is specifically configured to take a topic containing characters in a target topic in an image as a target topic to be selected; and determining the target topic according to the target topic to be selected.
In a possible design, the processing module 1302 is specifically configured to, if a target topic to be selected is one, take the target topic to be selected as a target topic; if the target questions to be selected are multiple, displaying the multiple target questions to be selected; and determining a target title according to the selection operation of the user on a plurality of target titles to be selected.
In one possible design, the search result of the target object is at least one of: the answer of the target topic, the analysis of the target topic and the check result of the answer of the target topic.
In one possible design, the transceiver module 1301 is further configured to send the target object to the server and receive a search result of the target object from the server.
In one possible design, the transceiver module 1301 is further configured to send an image to a server and receive a search result of a target object from the server.
Fig. 14 is a schematic structural diagram of a device for obtaining search results provided by the present application. The apparatus for obtaining the search result in this embodiment may be the server in the foregoing embodiment. As shown in fig. 14, the apparatus 1400 for obtaining search results includes: a transceiver module 1401 and a processing module 1402.
The transceiver module 1401 is configured to receive an image from a terminal device, where the image is an image of a preset region, and the preset region includes a target object.
The processing module 1402 is configured to obtain a search result of the target object according to the image.
The transceiver module 1401 is further configured to send a search result of the target object to the terminal device.
In one possible design, the processing module 1402 is specifically configured to obtain a target object in an image; and searching a search result of the target object in the database according to the target object.
In one possible design, the image includes a user's finger.
A processing module 1402, specifically configured to obtain a position of a fingertip of the user in the image; an object at the position of the user's fingertip is taken as a target object.
In one possible design, the target object is a target topic, and at least one topic is included in the image.
The processing module 1402 is specifically configured to identify a coverage area of each topic in the image according to an identifier of each topic in the image; and taking the title corresponding to the coverage range to which the position of the fingertip of the user belongs as a target title.
In one possible design, the search result of the target object is at least one of: the answer of the target topic, the analysis of the target topic and the check result of the answer of the target topic.
In another scenario, the transceiver module 1401 is configured to receive a target object from a terminal device.
The processing module 1402 is further configured to search the database for a search result of the target object according to the target object.
The transceiver module 1401 is further configured to send a search result of the target object to the terminal device.
The principle and technical effect of the apparatus for obtaining search results provided in this embodiment are similar to those of the method for obtaining search results, and are not described herein again.
According to an embodiment of the present application, an electronic device and a readable storage medium are also provided. Fig. 15 is a schematic structural diagram of an electronic device provided in the present application. The electronic device is the terminal device or the server in the above embodiment.
As shown in fig. 15, the electronic apparatus includes: one or more processors 1501, memory 1502, and interfaces for connecting the various components, including high-speed interfaces and low-speed interfaces. The various components are interconnected using different buses and may be mounted on a common motherboard or in other manners as desired. The processor may process instructions for execution within the electronic device, including instructions stored in or on the memory to display graphical information of a GUI on an external input/output apparatus (such as a display device coupled to the interface). In other embodiments, multiple processors and/or multiple buses may be used, along with multiple memories and multiple memories, as desired. Also, multiple electronic devices may be connected, with each device providing portions of the necessary operations (e.g., as a server array, a group of blade servers, or a multi-processor system). Fig. 15 illustrates an example of a processor 1501.
The memory 1502 is a non-transitory computer readable storage medium provided herein. Wherein the memory stores instructions executable by at least one processor to cause the at least one processor to perform the method of obtaining search results provided herein. The non-transitory computer readable storage medium of the present application stores computer instructions for causing a computer to perform the method of obtaining search results provided herein.
The memory 1502, which is a non-transitory computer readable storage medium, may be used to store non-transitory software programs, non-transitory computer executable programs, and modules, such as program instructions/modules corresponding to the method of obtaining search results in the embodiments of the present application. The processor 1501 executes various functional applications of the server and data processing, i.e., a method of acquiring search results in the above-described method embodiments, by running non-transitory software programs, instructions, and modules stored in the memory 1502.
The memory 1502 may include a program storage area that may store an operating system, an application program required for at least one function, and a data storage area; the storage data area may store data created according to use of the electronic device, and the like. Further, the memory 1502 may include high speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, the memory 1502 may optionally include memory located remotely from the processor 1501, which may be connected to the electronic device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The electronic device of the method of obtaining search results may further include: an input device 1503, an output device 1504, and a camera 1505. The processor 1501, the memory 1502, the input device 1503, the output device 1504, and the camera 1505 may be connected by a bus or other means, and are exemplified by a bus in fig. 15. The camera 1505 may capture an image containing a target object.
The input device 1503 may receive input numeric or character information and generate key signal inputs related to user settings and function control of the electronic device, such as a touch screen, a keypad, a mouse, a track pad, a touch pad, a pointing stick, one or more mouse buttons, a track ball, a joystick, or other input devices. The output devices 1504 may include a display device, auxiliary lighting devices (e.g., LEDs), and haptic feedback devices (e.g., vibrating motors), among others. The display device may include, but is not limited to, a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display, and a plasma display. In some implementations, the display device can be a touch screen.
Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, application specific ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
These computer programs (also known as programs, software applications, or code) include machine instructions for a programmable processor, and may be implemented using high-level procedural and/or object-oriented programming languages, and/or assembly/machine languages. As used herein, the terms "machine-readable medium" and "computer-readable medium" refer to any computer program product, apparatus, and/or device (e.g., magnetic discs, optical disks, memory, Programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term "machine-readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present application may be executed in parallel, sequentially, or in different orders, as long as the desired results of the technical solutions disclosed in the present application can be achieved, and the present invention is not limited herein.
The above-described embodiments should not be construed as limiting the scope of the present application. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (26)

1. A method of obtaining search results, comprising:
receiving a voice instruction of a user, wherein the voice instruction instructs to acquire a search result of a target object in a preset area;
acquiring an image of the preset area;
and acquiring a search result of the target object according to the image.
2. The method of claim 1, wherein obtaining the search result of the target object from the image comprises:
acquiring the target object in the image;
and acquiring a search result of the target object according to the target object.
3. The method of claim 2, wherein said acquiring the target object in the image comprises:
acquiring the position of the target object indicated by the user in an image;
and taking the object at the position as the target object.
4. The method of claim 3, wherein the image includes a finger of the user, and wherein the obtaining the position of the target object indicated by the user in the image includes:
acquiring the position of the fingertip of the user in the image;
and taking the position of the fingertip of the user as the position of the target object in the image.
5. The method of claim 3, wherein the voice instruction indicates a position of the target object in an image, and wherein the obtaining the position of the target object in the image indicated by the user comprises:
and determining the position of the target object in the image according to the voice instruction.
6. The method according to claim 4 or 5, wherein the target object is a target topic, at least one topic is included in the image, and the taking the object at the position as the target object comprises:
identifying the coverage range of each topic in the image according to the identification of each topic in the image;
and taking the title corresponding to the coverage range to which the position belongs as the target title.
7. The method of claim 2, wherein the voice instruction indicates the target object, and wherein the capturing the target object in the image comprises:
and acquiring the target object in the image according to the voice instruction.
8. The method according to claim 7, wherein the target object is a target topic, the voice command includes an identifier of the target topic or a text in the target topic, and the obtaining the target object in the image includes:
identifying the target title and target identification or characters in the target title in the voice command;
and acquiring the target title in the image according to the target title identification or the characters in the target title.
9. The method according to claim 8, wherein the image includes at least one topic, and the obtaining the target topic in the image according to the identification of the target topic includes:
identifying an identity of each topic in the image;
and taking the topic with the same identification as the target topic.
10. The method according to claim 8 or 9, wherein the obtaining the target topic in the image according to the text in the target topic comprises:
taking the title of the characters contained in the target title in the image as a target title to be selected;
and determining the target topic according to the target topic to be selected.
11. The method according to claim 10, wherein the determining the target topic according to the target topic to be selected comprises:
if the target question to be selected is one, taking the target question to be selected as the target question;
if the target questions to be selected are multiple, displaying the multiple target questions to be selected;
and determining the target title according to the selection operation of the user on a plurality of target titles to be selected.
12. The method according to any one of claims 6, 8-11,
the search result of the target object is at least one of the following: the answer of the target subject, the analysis of the target subject and the checking result of the answer of the target subject.
13. The method according to claim 2, wherein the obtaining the search result of the target object according to the target object comprises:
sending the target object to a server;
receiving a search result of the target object from the server.
14. The method of claim 1, wherein obtaining the search result of the target object from the image comprises:
sending the image to a server;
receiving a search result of the target object from the server.
15. A method of obtaining search results, comprising:
receiving an image from a terminal device, wherein the image is an image of a preset area, and the preset area comprises a target object;
obtaining a search result of the target object according to the image;
and sending the search result of the target object to the terminal equipment.
16. The method of claim 15, wherein obtaining the search result of the target object from the image comprises:
acquiring the target object in the image;
and searching a database for the search result of the target object according to the target object.
17. The method of claim 16, wherein the image includes a finger of a user, and wherein acquiring the target object in the image includes:
acquiring the position of the fingertip of the user in the image;
taking an object at a position of the user fingertip as the target object.
18. The method of claim 17, wherein the target object is a target topic, wherein at least one topic is included in the image, and wherein the taking an object at a position of the user fingertip as the target object comprises:
identifying the coverage range of each topic in the image according to the identification of each topic in the image;
and taking the title corresponding to the coverage range to which the position of the fingertip of the user belongs as the target title.
19. The method of claim 18,
the search result of the target object is at least one of the following: the answer of the target subject, the analysis of the target subject and the checking result of the answer of the target subject.
20. A method of obtaining search results, comprising:
receiving a target object from a terminal device;
searching a database for a search result of the target object according to the target object;
and sending the search result of the target object to the terminal equipment.
21. The method of claim 20, wherein the target object is a target topic, and wherein the search result of the target object is at least one of: the answer of the target subject, the analysis of the target subject and the checking result of the answer of the target subject.
22. An apparatus for obtaining search results, comprising:
the receiving and sending module is used for receiving a voice instruction of a user, and the voice instruction acquires a search result of a target object in a preset area;
and the processing module is used for acquiring the image of the preset area and acquiring the search result of the target object according to the image.
23. An apparatus for obtaining search results, comprising:
the receiving and sending module is used for receiving an image from the terminal equipment, wherein the image is an image of a preset area, and the preset area comprises a target object;
the processing module is used for acquiring a search result of the target object according to the image;
the transceiver module is further configured to send the search result of the target object to the terminal device.
24. An apparatus for obtaining search results, comprising:
the receiving and sending module is used for receiving the target object from the terminal equipment;
the processing module is used for searching a search result of the target object in a database according to the target object;
the transceiver module is further configured to send the search result of the target object to the terminal device.
25. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-21.
26. A non-transitory computer readable storage medium having stored thereon computer instructions for causing a computer to perform the method of any one of claims 1-21.
CN202010367988.0A 2020-04-30 2020-04-30 Method and device for obtaining search result, electronic equipment and storage medium Pending CN113590864A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010367988.0A CN113590864A (en) 2020-04-30 2020-04-30 Method and device for obtaining search result, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010367988.0A CN113590864A (en) 2020-04-30 2020-04-30 Method and device for obtaining search result, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN113590864A true CN113590864A (en) 2021-11-02

Family

ID=78237719

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010367988.0A Pending CN113590864A (en) 2020-04-30 2020-04-30 Method and device for obtaining search result, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113590864A (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105912717A (en) * 2016-04-29 2016-08-31 广东小天才科技有限公司 Image-based information searching method and device
CN107908801A (en) * 2017-12-25 2018-04-13 广东小天才科技有限公司 Question searching method based on voice and electronic equipment
CN108182270A (en) * 2018-01-17 2018-06-19 广东小天才科技有限公司 Search content transmission method, search content search method, smart pen, search terminal, and storage medium
CN109710750A (en) * 2019-01-23 2019-05-03 广东小天才科技有限公司 Question searching method and learning equipment
CN109753583A (en) * 2019-01-16 2019-05-14 广东小天才科技有限公司 Question searching method and electronic equipment
CN111027353A (en) * 2019-02-18 2020-04-17 广东小天才科技有限公司 Search content extraction method and electronic equipment
CN111026949A (en) * 2019-02-26 2020-04-17 广东小天才科技有限公司 Question searching method and system based on electronic equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105912717A (en) * 2016-04-29 2016-08-31 广东小天才科技有限公司 Image-based information searching method and device
CN107908801A (en) * 2017-12-25 2018-04-13 广东小天才科技有限公司 Question searching method based on voice and electronic equipment
CN108182270A (en) * 2018-01-17 2018-06-19 广东小天才科技有限公司 Search content transmission method, search content search method, smart pen, search terminal, and storage medium
CN109753583A (en) * 2019-01-16 2019-05-14 广东小天才科技有限公司 Question searching method and electronic equipment
CN109710750A (en) * 2019-01-23 2019-05-03 广东小天才科技有限公司 Question searching method and learning equipment
CN111027353A (en) * 2019-02-18 2020-04-17 广东小天才科技有限公司 Search content extraction method and electronic equipment
CN111026949A (en) * 2019-02-26 2020-04-17 广东小天才科技有限公司 Question searching method and system based on electronic equipment

Similar Documents

Publication Publication Date Title
EP3258423B1 (en) Handwriting recognition method and apparatus
US11288031B2 (en) Information processing apparatus, information processing method, and information processing system
CN110297550B (en) Label display method and device, screen throwing equipment, terminal and storage medium
CN105378817A (en) Incorporating external dynamic content into a whiteboard
US20110252316A1 (en) Translating text on a surface computing device
US9335965B2 (en) System and method for excerpt creation by designating a text segment using speech
JP2017191601A (en) Method and system for keyword search using messenger service
JP6434640B2 (en) Message display method, message display device, and message display device
US20220141532A1 (en) Techniques for rich interaction in remote live presentation and accurate suggestion for rehearsal through audience video analysis
US10990344B2 (en) Information processing apparatus, information processing system, and information processing method
CN102681870A (en) Automatically performing an action upon a login
CN102939574A (en) Character selection
US11620414B2 (en) Display apparatus, display method, and image processing system
US10965743B2 (en) Synchronized annotations in fixed digital documents
US10657326B2 (en) Removable spell checker device
US20170017632A1 (en) Methods and Systems of Annotating Local and Remote Display Screens
CN111242455A (en) Method and device for evaluating voice function of electronic map, electronic equipment and storage medium
CN111858880A (en) Method and device for obtaining query result, electronic equipment and readable storage medium
CN111638787A (en) Method and device for displaying information
US20230127120A1 (en) Machine learning driven teleprompter
CN113590864A (en) Method and device for obtaining search result, electronic equipment and storage medium
WO2023075909A1 (en) Machine learning driven teleprompter
CN113190365B (en) Information processing method and device and electronic equipment
CN115379113A (en) Shooting processing method, device, equipment and storage medium
CN112579868A (en) Multi-modal graph recognition searching method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination