US20140215382A1 - Method for Utilizing Projected Gesture Completion to Improve Instrument Performance - Google Patents

Method for Utilizing Projected Gesture Completion to Improve Instrument Performance Download PDF

Info

Publication number
US20140215382A1
US20140215382A1 US13/750,536 US201313750536A US2014215382A1 US 20140215382 A1 US20140215382 A1 US 20140215382A1 US 201313750536 A US201313750536 A US 201313750536A US 2014215382 A1 US2014215382 A1 US 2014215382A1
Authority
US
United States
Prior art keywords
gesture
processing system
data processing
screen
completed
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/750,536
Inventor
Glenn R. Engel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Keysight Technologies Inc
Original Assignee
Agilent Technologies Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Agilent Technologies Inc filed Critical Agilent Technologies Inc
Priority to US13/750,536 priority Critical patent/US20140215382A1/en
Assigned to AGILENT TECHNOLOGIES, INC. reassignment AGILENT TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ENGEL, GLENN R.
Priority to CN201310724114.6A priority patent/CN103970439A/en
Priority to CN202210128082.2A priority patent/CN114779921A/en
Publication of US20140215382A1 publication Critical patent/US20140215382A1/en
Assigned to KEYSIGHT TECHNOLOGIES, INC. reassignment KEYSIGHT TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AGILENT TECHNOLOGIES, INC.
Priority to US17/035,610 priority patent/US11379114B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning

Definitions

  • gestures on a touch display screen to control an underlying computer program has become commonplace.
  • There is a class of gestures in which the user moves one or more fingers while in continuous contact with the screen. For example, the user may scroll down a list or pan a view by touching the screen and moving a finger in the desired direction.
  • pinch gestures are used to zoom in or out on a scene that is displayed on the screen.
  • Touch screens are also being used to control measurement instruments.
  • the screen can be the display on a computer that directly controls the measurements or on a mobile device that is connected to the instrument via a network that may include the Internet and/or part of a wireless telecommunication network.
  • a mobile device is being used to control an instrument. The control depends on the data that is currently displayed on the screen of the mobile device. To update the screen data, the mobile device must send the appropriate command to the instrument and wait for the instrument to return the new data. If the time to execute the gesture is a significant fraction of the total time to update the display, the gesture introduces a delay into the control of the instrument and/or the display of information on the screen. Such delays lead to an interface that is “jumpy” or “sluggish”. If the gesture is part of a control loop, the resulting system can oscillate.
  • the user may use a “fling” gesture. If the mobile device waits until the gesture is completed, there is a delay between starting the gesture and viewing the new data on the screen. This delay is added to any delay introduced by the instrument and the communication link in returning the data indicated by the gesture. If the user determines that further scrolling is needed, another gesture is executed, and so on.
  • the present invention includes a method of operating a first data processing system having a screen for responding to a vector gesture performed on that screen.
  • the method includes estimating a direction and assuming an identity for the vector gesture prior to the vector gesture being completed.
  • a command that depends on the estimated direction and assumed identity is executed, the command being a command that would be required to update a current state of the first data processing system if the assumed identity of the vector gesture is correct.
  • the command is executed prior to the vector gesture being completed.
  • the current state of the first data processing system is then updated based on the data generated by the command.
  • the first data processing system is updated to a state indicated by the completed vector gesture.
  • the command causes a second data processing system to provide data to the first data processing system.
  • updating the current state of the first data processing system includes updating a display on the first data processing system with data obtained from executing the command.
  • the current state includes a first scene on the screen, and the assumed identity includes a scrolling gesture that indicates that the first scene is to be scrolled in the determined direction.
  • the current state includes displaying a first scene on the screen, and updating the current state includes providing a partially scrolled version of the first scene.
  • updating the first data processing system to a state indicated by the completed vector gesture includes displaying a first scene on the screen followed by a second scene having a higher resolution than the first scene.
  • estimating a direction and assuming an identity for the vector gesture are repeated if the vector gesture is not completed when the results of the command are completed.
  • FIG. 1 illustrates a mobile device that utilizes the method of the present invention.
  • FIG. 2 illustrates the manner in which a zoom out gesture is implemented in one embodiment of the present invention.
  • FIGS. 4-6 illustrate a zoom in operation in which the missing data points are sent in an order that allows an approximation to the final display to be generated before the entire data set needed for the display is sent from the data processor.
  • FIG. 7 is a flow chart for one embodiment of a method according to the present invention.
  • the present invention can be more easily understood with regard to a class of gestures that involve the movement of an object on the screen in which the gesture involves an object that moves in a first direction and ends either by slowing the rate of motion or the lifting of the object off of the screen.
  • the object is typically one or more fingers of a user.
  • the teachings of the present invention can be applied to other objects such as a stylus.
  • the object could be a virtual object such as the pointer controlled by a mouse. The pointer “touches” the screen when a specific button is depressed and leaves the screen when the button is released.
  • the direction of motion is typically a linear direction; however, gestures that utilize rotational movements can also be accommodated. For example, a gesture with two fingers could simulate turning a knob that is displayed on the screen.
  • the present invention is based on the observation that, in many cases, the operations that must be performed to provide an updated display that is specified through a gesture can be started, if not completed, before the gesture is completed. Hence, the display can be updated while the gesture is being performed to provide a display that is more nearly consistent with the gesture at any point during the performance of the gesture.
  • FIG. 1 illustrates a mobile device that utilizes the method of the present invention.
  • Mobile device 20 has a touch-enabled screen 22 that displays data to the user of mobile device 20 and receives commands in the form of touch gestures.
  • Mobile device 20 communicates with data processor 31 that stores or generates data that is displayed on mobile device 20 in response to touch gestures performed on screen 22 .
  • data processor 31 stores or generates data that is displayed on mobile device 20 in response to touch gestures performed on screen 22 .
  • mobile device 20 displays a portion of a list 33 that is received from data processor 31 .
  • Screen 22 can be viewed as a window through which the user views a portion of the list, which is too large to be able to be seen all at once.
  • the user of mobile device 20 wishes to scroll through list 33 by executing a fling gesture on screen 22 .
  • a fling gesture is indicated by the user moving the user's finger 34 on screen 22 while maintaining contact with screen 22 in a rapid motion in which the finger is still moving when the finger is lifted.
  • the list scrolls through the window in a manner in which the user sees the portion of the list currently in the window.
  • the speed at which the list scrolls through the window after the user lifts his or her finger decreases over time until the scrolling finally stops leaving a portion of the list visible in the window if the user takes no further action.
  • the direction in which the window is to move relative to list 33 can be determined as soon as the user begins the gesture. However, the distance that mobile device 20 is to shift the list will not be known until the gesture is completed.
  • mobile device 20 determines the direction and signals data processor 31 to begin generating and sending the data in region 35 of list 33 .
  • the data is sent in the order the data appears in the list with the data nearest the portion of the list that is currently being displayed sent first.
  • mobile device 20 updates screen 22 by augmenting the portion of the list that is visible in the indicated direction. Thus, the update of the display commences prior to the gesture being completed.
  • mobile device 20 When the gesture is completed, mobile device 20 ascertains the extent of the requested scroll. If the updated display already reflects the desired scroll, mobile device 20 terminates the update operation. However, in general, the displayed portion of the list will not coincide with the requested update as determined by the fling operation. If the amount of data received from data processor 31 at the end of the gesture is less than that needed to satisfy the gesture, mobile device 20 continues to request data from processor 31 until the desired data has been received and displayed. If the data received at the end of the gesture is sufficient to update the display, mobile device 20 adjusts the display to reflect the indicated amount of scrolling. This may involve rolling back the scrolled list if the display has already been scrolled past the desired end point.
  • these intermediate list items can be sent and displayed in a low bandwidth form that is sufficient for the user to determine if the user wants to stop the scroll.
  • a low bandwidth form of the list is defined to be a form that can be sent to the mobile device in less time than the list would be sent if the list were to be displayed on the screen in final form.
  • this could be a form in which only the first few words of each list item are sent, a form in which the items are sent with abbreviations in place of the words or phrases, or a form in which every n th entry in the list is sent.
  • the correct form of each item that is to be displayed is used in place of the low bandwidth form.
  • the use of the reduced bandwidth form of the list frees bandwidth for sending the final form of the portion of the list that is to be displayed or a more esthetic scroll that moves at the desired speed.
  • the update information is sent in an order that allows the user to view the data near the end of the update prior to the other information appearing. For example, if the last entry in the updated list is known, that entry can be sent next so that it appears on mobile device 20 .
  • the missing items are then sent in an order that depends on the particular mode of operation. For example, the missing items can be sent in a random order so that the user can view items at different locations in the update list to determine if further scrolling is required.
  • the list can be filled by working backwards from the last entry or forward from the last received entry that is in the completed portion of the list.
  • This type of non-ordered update can be more easily understood with respect to a display that shows an analog signal such as a graph of the signal intensity as a function of time or some other parameter such as frequency.
  • Such displays are common in mobile devices that control or view data from an instrument such as an oscilloscope or frequency analyzer.
  • a scene that shows a signal as a function of some variable, X.
  • the user controls the zoom level of the signal to alter the portion and resolution of the signal displayed on a device with a touch screen by executing a “zoom gesture”.
  • zoom gesture For the purposes of this example, it will be assumed that the user wishes to zoom out thereby displaying more of the signal at a lower resolution on a mobile device that is used to control the instrument that provides the data.
  • data related to the part of the scene that is not currently being displayed is required.
  • the mobile device determines that a zoom out has started, the mobile device can begin to acquire additional data on each side of the current scene and add that data to the current display.
  • the computer can predict approximately where the gesture will complete. For example, if a pinch zoom gesture is used, the pinch zoom cannot exceed the width of the user's finger span; hence, the maximum amount of data that may be required can be predicted when the pinch zoom begins.
  • FIG. 2 illustrates the manner in which a zoom out gesture is implemented in one embodiment of the present invention.
  • a touch-enabled display screen 12 that is part of a mobile device 10 currently shows the portion of curve 13 that is within the boundaries of display screen 12 .
  • a user 14 of mobile device 10 signals data processor 11 , which either stores the entire trace or can generate the portions that are outside the range currently displayed by mobile device 10 , by performing a pinch gesture on display screen 12 .
  • the degree of the zoom is determined by the distance that the user moves the user's fingers together during the pinch zoom gesture. This information will not be available until the user completes the gesture.
  • mobile device 10 signals data processor 11 to begin sending the needed portions outside the current window such as portion 15 .
  • the data is sent in an order that gives priority to the portions of curve 13 that are nearest the portions already being displayed, since those are the portions that are most likely going to be needed to satisfy the pinch zoom command.
  • Data processor 11 begins sending the portions that may be needed before the pinch gesture is completed.
  • the display is updated by decimating the portion of the curve that is to remain displayed at the lower resolution and adding the new data to that portion as the new data is sent.
  • the actual data that is sent to mobile device 10 by the time the pinch gesture is completed may be more or less than the data that is needed to satisfy the pinch gesture.
  • mobile device 10 continuously updates display screen 12 with the data as the data arrives. If the data available at mobile device 10 is sufficient to generate the correct display if the pinch zoom operation were to terminate at its current location, the display will smoothly follow the pinch zoom. If the data is not sufficient, the display will appear to be continuously changing and will catch up after the pinch zoom terminates. Mobile device 10 can determine the maximum amount of data needed to satisfy the pinch zoom, since that corresponds to the minimum distance between the user's fingers. Hence, if data is sent faster than needed to keep up with the pinch zoom, at most, some additional data will have been sent to mobile device 10 if the gesture ends at a point less than this minimum distance.
  • the missing portions of the curve can be provided in an order that allows an approximation to the final display to be generated by the time the pinch zoom gesture is completed.
  • the approximation is then refined as more data is received.
  • an approximation to the final curve could be generated by sending every nth point to mobile device 10 in the region in which data is missing. These points would then be fitted to a curve that is displayed on display screen 12 . After these points have been sent, points that are intermediate to these would then be sent.
  • the data points are then refitted to a curve and displayed.
  • the display appears at its final resolution and updates provide a more detailed view of the data. Since the human eye has difficulty perceiving rapidly changing scenes at high resolution, this form of update hides the underlying “noise” resulting from the approximation of the curve by the incomplete data set.
  • the missing data is the data between the existing points on the portion of the display that remains within the region displayed on display screen 12 .
  • FIGS. 4-6 which illustrate a zoom in operation in which the missing data points are sent in an order that allows an approximation to the final display to be generated before the entire data set needed for the display is sent from the data processor.
  • user 14 indicates that the user wishes to zoom in on the portion of curve 53 shown at 54 by executing a zoom in pinch gesture.
  • the final display will include some of the data points that are currently stored in mobile device 10 together with a set of additional points that provide information about curve 53 between the points that are currently being displayed.
  • mobile device 10 When the zoom in operation commences, mobile device 10 signals data processor 11 to commence sending data points between the existing data points. By the time the zoom in gesture is completed, mobile device 10 has accumulated the data points in FIG. 5 . Exemplary data points are labeled at 41 - 45 . These data points are used to generate an approximation 49 to the final display. As additional data points arrive for locations between the existing data point locations such as data point 46 shown in FIG. 6 , display screen 12 is updated to provide a better approximation to the final curve as shown at 51 .
  • a vector touch gesture is defined to be a gesture that is characterized by a direction and a magnitude in which the exact magnitude is not known until the gesture is completed.
  • the direction in contrast, can be determined by the data processing system prior to the time at which the magnitude can be ascertained.
  • FIG. 7 is a flow chart for one embodiment of a method according to the present invention.
  • the data processing system that operates the touch-enabled display detects a touch gesture that is initiated on the touch-enabled display and determines if the gesture is a vector gesture as shown at 71 . If the touch gesture is a vector touch gesture, the data processing system determines the direction associated with that touch direction as shown at 72 . In addition, the data processing system determines a set of vector touch gestures that could be the one being implemented as shown at 73 .
  • a scrolling operation could be performed by a fling or a conventional scroll in which the user moves the user's finger in a continuous manner while in contact with the screen but stops the motion before lifting the finger.
  • a fling in contrast, is characterized by the finger being in motion at a speed greater than some threshold speed when the finger is lifted. The data processing system cannot determine which scrolling operation is being requested until the user actually lifts the user's finger.
  • the data processing system uses the determined direction and the current state of the gesture to execute a command that will at least partially perform work needed to respond to one or more of the possible gestures as shown at 74 .
  • the data processing system uses the determined direction and the current state of the gesture to execute a command that will at least partially perform work needed to respond to one or more of the possible gestures as shown at 74 .
  • the direction is determined, the finger position or positions will have advanced to a new location that provides some additional information about the magnitude associated with the vector gesture. For example, from the current position, the data processing system can determine the minimum distance that the display will need to be scrolled, and hence, the minimum amount of new information needed to execute the scroll.
  • a maximum distance the display could be scrolled in the current possible gesture can also be determined. For example, if the current gesture were a conventional scroll, the maximum distance is determined by measuring the scroll distance corresponding to the user's finger reaching the edge of the screen in the current direction. If the scroll operation was a fling, the maximum distance would be determined by the current scroll speed and some estimate for the maximum speed of the finger when it is lifted.
  • the command returns data needed to provide the display in a scrolled state that is a better approximation to the final state than the current state.
  • the system is updated based on the results of those commands as shown at 75 .
  • an updated display could be generated based on the results of the executed commands as the data from those commands is received.
  • the commands can continue to elicit data needed to execute the maximum possible scroll in the case of a scroll. If the scroll requires less than this amount of data, the excess data can be ignored or saved to facilitate a future operation.
  • the data processing system determines if the gesture has been completed as shown at 76 . If the gesture has not been completed, the current state of the gesture is determined as shown at 77 . The state is determined from the current locations of the user's fingers. This information is used to update the possible set of gestures and determine further commands to be executed based on the current estimate of the fingers.
  • the data processing completes the operations specified by the gesture as shown at 78 .
  • the completion may require that the results from one or more commands that were previously issued be received.
  • the completion may require additional commands. For example, in the case of a fling gesture, additional data may need to be requested before the screen can be updated to its final form.
  • actions that control a device that is in communication with the data processing system could also be utilized.
  • the command could be a command that is sent to an external device that causes that device to carry out an operation.
  • the operation in question could result in an updated display; however, such an update is not necessarily required.
  • an instrument that is separate from the mobile device may be updated without changing the display on the mobile device.
  • the above-described embodiments have utilized a mobile device with a touch-enabled screen.
  • the present invention can be practiced on any data processing system that has a touch-enabled screen and software to detect one or more vector gestures.
  • Touch-enabled screens are available on desktop systems as well as mobile devices. It should also be noted that the screen view being updated need not be the touch-enabled screen.
  • a system with two screens, one touch enabled and the other a conventional screen could benefit from the present invention being utilized to receive the vector gestures on the touch-enabled screen, while updating the conventional display.
  • the present invention also includes any computer readable medium that stores instructions that cause a computer reading that medium to execute the method of the present invention.
  • any computer readable medium that qualifies as patentable subject matter under 35 U.S.C. 101 will be defined to be a computer readable storage medium.
  • Such media include non-transitory storage media such as non-volatile memories, magnetic and optically-based disk drives and storage elements read by those disk drives.

Abstract

A method of operating a first data processing system having a screen for responding to a vector gesture performed on that screen is disclosed. The method includes estimating a direction and assuming an identity for the vector gesture prior to the vector gesture being completed. A command that depends on the estimated direction and assumed identity is executed, the command being a command that would be required to update a current state of the first data processing system if the assumed identity of the vector gesture is correct. The command is executed prior to the vector gesture being completed. The current state of the first data processing system is then updated based on the data generated by the command. After the vector gesture has been determined to have been completed, the first data processing system is updated to a state indicated by the completed vector gesture.

Description

    BACKGROUND
  • The use of gestures on a touch display screen to control an underlying computer program has become commonplace. There is a class of gestures in which the user moves one or more fingers while in continuous contact with the screen. For example, the user may scroll down a list or pan a view by touching the screen and moving a finger in the desired direction. Similarly, “pinch” gestures are used to zoom in or out on a scene that is displayed on the screen.
  • Touch screens are also being used to control measurement instruments. The screen can be the display on a computer that directly controls the measurements or on a mobile device that is connected to the instrument via a network that may include the Internet and/or part of a wireless telecommunication network. Consider a case in which a mobile device is being used to control an instrument. The control depends on the data that is currently displayed on the screen of the mobile device. To update the screen data, the mobile device must send the appropriate command to the instrument and wait for the instrument to return the new data. If the time to execute the gesture is a significant fraction of the total time to update the display, the gesture introduces a delay into the control of the instrument and/or the display of information on the screen. Such delays lead to an interface that is “jumpy” or “sluggish”. If the gesture is part of a control loop, the resulting system can oscillate.
  • For example, if the user is looking for a value in a list by scrolling through the list, the user may use a “fling” gesture. If the mobile device waits until the gesture is completed, there is a delay between starting the gesture and viewing the new data on the screen. This delay is added to any delay introduced by the instrument and the communication link in returning the data indicated by the gesture. If the user determines that further scrolling is needed, another gesture is executed, and so on.
  • SUMMARY
  • The present invention includes a method of operating a first data processing system having a screen for responding to a vector gesture performed on that screen. The method includes estimating a direction and assuming an identity for the vector gesture prior to the vector gesture being completed. A command that depends on the estimated direction and assumed identity is executed, the command being a command that would be required to update a current state of the first data processing system if the assumed identity of the vector gesture is correct. The command is executed prior to the vector gesture being completed. The current state of the first data processing system is then updated based on the data generated by the command. After the vector gesture has been determined to have been completed, the first data processing system is updated to a state indicated by the completed vector gesture.
  • In one aspect of the invention, the command causes a second data processing system to provide data to the first data processing system.
  • In another aspect of the invention, updating the current state of the first data processing system includes updating a display on the first data processing system with data obtained from executing the command. In a still further aspect of the invention, the current state includes a first scene on the screen, and the assumed identity includes a scrolling gesture that indicates that the first scene is to be scrolled in the determined direction.
  • In a further aspect of the invention, the current state includes displaying a first scene on the screen, and updating the current state includes providing a partially scrolled version of the first scene.
  • In another aspect of the invention, updating the first data processing system to a state indicated by the completed vector gesture includes displaying a first scene on the screen followed by a second scene having a higher resolution than the first scene.
  • In a further aspect of the invention, estimating a direction and assuming an identity for the vector gesture are repeated if the vector gesture is not completed when the results of the command are completed.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a mobile device that utilizes the method of the present invention.
  • FIG. 2 illustrates the manner in which a zoom out gesture is implemented in one embodiment of the present invention.
  • FIGS. 4-6 illustrate a zoom in operation in which the missing data points are sent in an order that allows an approximation to the final display to be generated before the entire data set needed for the display is sent from the data processor.
  • FIG. 7 is a flow chart for one embodiment of a method according to the present invention.
  • DETAILED DESCRIPTION
  • The present invention can be more easily understood with regard to a class of gestures that involve the movement of an object on the screen in which the gesture involves an object that moves in a first direction and ends either by slowing the rate of motion or the lifting of the object off of the screen. The object is typically one or more fingers of a user. However, the teachings of the present invention can be applied to other objects such as a stylus. It should also be noted that the object could be a virtual object such as the pointer controlled by a mouse. The pointer “touches” the screen when a specific button is depressed and leaves the screen when the button is released.
  • The direction of motion is typically a linear direction; however, gestures that utilize rotational movements can also be accommodated. For example, a gesture with two fingers could simulate turning a knob that is displayed on the screen.
  • The present invention is based on the observation that, in many cases, the operations that must be performed to provide an updated display that is specified through a gesture can be started, if not completed, before the gesture is completed. Hence, the display can be updated while the gesture is being performed to provide a display that is more nearly consistent with the gesture at any point during the performance of the gesture.
  • The manner in which the present invention provides its advantages can be more easily understood with reference to FIG. 1, which illustrates a mobile device that utilizes the method of the present invention. Mobile device 20 has a touch-enabled screen 22 that displays data to the user of mobile device 20 and receives commands in the form of touch gestures. Mobile device 20 communicates with data processor 31 that stores or generates data that is displayed on mobile device 20 in response to touch gestures performed on screen 22. In the example shown in FIG. 1, mobile device 20 displays a portion of a list 33 that is received from data processor 31. Screen 22 can be viewed as a window through which the user views a portion of the list, which is too large to be able to be seen all at once.
  • The user of mobile device 20 wishes to scroll through list 33 by executing a fling gesture on screen 22. For the purposes of this discussion, a fling gesture is indicated by the user moving the user's finger 34 on screen 22 while maintaining contact with screen 22 in a rapid motion in which the finger is still moving when the finger is lifted. During the fling gesture, the list scrolls through the window in a manner in which the user sees the portion of the list currently in the window. The speed at which the list scrolls through the window after the user lifts his or her finger decreases over time until the scrolling finally stops leaving a portion of the list visible in the window if the user takes no further action. However, if the user touches the screen again after lifting his or her finger and during the time the list is still scrolling, the scrolling stops leaving the currently visible portion of the list in the window. Hence, the scroll gesture is not completed until the scrolling stops, since the gesture could include a premature stop indicated by the user again touching the screen. The direction of the motion of the user's finger specifies which way the list is to be scrolled. The speed with which the user moves finger 34 determines the time and speed after the lifting of the finger during which the list will continue to scroll. Data processor 31 sends the data needed to augment the display on mobile device 20 in response to the gesture.
  • The direction in which the window is to move relative to list 33 can be determined as soon as the user begins the gesture. However, the distance that mobile device 20 is to shift the list will not be known until the gesture is completed. In the present invention, mobile device 20 determines the direction and signals data processor 31 to begin generating and sending the data in region 35 of list 33. In one aspect of the invention, the data is sent in the order the data appears in the list with the data nearest the portion of the list that is currently being displayed sent first. As the data is received, mobile device 20 updates screen 22 by augmenting the portion of the list that is visible in the indicated direction. Thus, the update of the display commences prior to the gesture being completed.
  • When the gesture is completed, mobile device 20 ascertains the extent of the requested scroll. If the updated display already reflects the desired scroll, mobile device 20 terminates the update operation. However, in general, the displayed portion of the list will not coincide with the requested update as determined by the fling operation. If the amount of data received from data processor 31 at the end of the gesture is less than that needed to satisfy the gesture, mobile device 20 continues to request data from processor 31 until the desired data has been received and displayed. If the data received at the end of the gesture is sufficient to update the display, mobile device 20 adjusts the display to reflect the indicated amount of scrolling. This may involve rolling back the scrolled list if the display has already been scrolled past the desired end point.
  • Consider the case in which data processor 31 cannot provide the information to mobile device 20 with sufficient speed to allow mobile device 20 to provide the user with the completed scrolled display at the end of the gesture. This situation can be the result of the limitations in the bandwidth between data processor 31 and mobile device 20, as well as limitations in the time needed to generate the data in data processor 31. If screen 22 is updated as the information arrives and the information is sent in the order the data appears in the list, then screen 22 will show a continuous slow update, which may not be the most esthetic form of update from the user's point of view, since the user may be more interested in the data near the end of the updated portion of the list.
  • Consider the case of a fling. When the user lifts his or her finger from the display, the end point of the list that will be in the final screen display is known assuming that the user does not stop the fling prematurely by touching the screen. However, to provide the user with the view of the data that scrolls through the window on the list on the way to the final screen display, the various intermediate list items must be sent and displayed to allow the user to decide if the user wishes to prematurely end the scroll.
  • In one aspect of the invention, these intermediate list items can be sent and displayed in a low bandwidth form that is sufficient for the user to determine if the user wants to stop the scroll. For the purposes of this discussion, a low bandwidth form of the list is defined to be a form that can be sent to the mobile device in less time than the list would be sent if the list were to be displayed on the screen in final form. For example, this could be a form in which only the first few words of each list item are sent, a form in which the items are sent with abbreviations in place of the words or phrases, or a form in which every nth entry in the list is sent. If the user stops the scroll prematurely based on this low bandwidth form of the list, the correct form of each item that is to be displayed is used in place of the low bandwidth form. The use of the reduced bandwidth form of the list frees bandwidth for sending the final form of the portion of the list that is to be displayed or a more esthetic scroll that moves at the desired speed.
  • In another aspect of the invention, the update information is sent in an order that allows the user to view the data near the end of the update prior to the other information appearing. For example, if the last entry in the updated list is known, that entry can be sent next so that it appears on mobile device 20. The missing items are then sent in an order that depends on the particular mode of operation. For example, the missing items can be sent in a random order so that the user can view items at different locations in the update list to determine if further scrolling is required. Alternatively, the list can be filled by working backwards from the last entry or forward from the last received entry that is in the completed portion of the list.
  • This type of non-ordered update can be more easily understood with respect to a display that shows an analog signal such as a graph of the signal intensity as a function of time or some other parameter such as frequency. Such displays are common in mobile devices that control or view data from an instrument such as an oscilloscope or frequency analyzer.
  • For example, consider a scene that shows a signal as a function of some variable, X. The user controls the zoom level of the signal to alter the portion and resolution of the signal displayed on a device with a touch screen by executing a “zoom gesture”. For the purposes of this example, it will be assumed that the user wishes to zoom out thereby displaying more of the signal at a lower resolution on a mobile device that is used to control the instrument that provides the data. To provide the new display, data related to the part of the scene that is not currently being displayed is required. Once the mobile device determines that a zoom out has started, the mobile device can begin to acquire additional data on each side of the current scene and add that data to the current display. When the user finishes the zoom operation, the additional data will already have been provided, or at least some significant fraction thereof. This provides the appearance of a display that is continuously updated with the zoom operation. In one aspect of the invention, the computer can predict approximately where the gesture will complete. For example, if a pinch zoom gesture is used, the pinch zoom cannot exceed the width of the user's finger span; hence, the maximum amount of data that may be required can be predicted when the pinch zoom begins.
  • Refer now to FIG. 2, which illustrates the manner in which a zoom out gesture is implemented in one embodiment of the present invention. A touch-enabled display screen 12 that is part of a mobile device 10 currently shows the portion of curve 13 that is within the boundaries of display screen 12. A user 14 of mobile device 10, signals data processor 11, which either stores the entire trace or can generate the portions that are outside the range currently displayed by mobile device 10, by performing a pinch gesture on display screen 12. The degree of the zoom is determined by the distance that the user moves the user's fingers together during the pinch zoom gesture. This information will not be available until the user completes the gesture.
  • However, as soon mobile device 10 recognizes the gesture as a pinch zoom gesture, mobile device 10 signals data processor 11 to begin sending the needed portions outside the current window such as portion 15. As noted above, in one aspect of the invention, the data is sent in an order that gives priority to the portions of curve 13 that are nearest the portions already being displayed, since those are the portions that are most likely going to be needed to satisfy the pinch zoom command. Data processor 11 begins sending the portions that may be needed before the pinch gesture is completed. The display is updated by decimating the portion of the curve that is to remain displayed at the lower resolution and adding the new data to that portion as the new data is sent.
  • The actual data that is sent to mobile device 10 by the time the pinch gesture is completed may be more or less than the data that is needed to satisfy the pinch gesture. During the execution of the gesture, mobile device 10 continuously updates display screen 12 with the data as the data arrives. If the data available at mobile device 10 is sufficient to generate the correct display if the pinch zoom operation were to terminate at its current location, the display will smoothly follow the pinch zoom. If the data is not sufficient, the display will appear to be continuously changing and will catch up after the pinch zoom terminates. Mobile device 10 can determine the maximum amount of data needed to satisfy the pinch zoom, since that corresponds to the minimum distance between the user's fingers. Hence, if data is sent faster than needed to keep up with the pinch zoom, at most, some additional data will have been sent to mobile device 10 if the gesture ends at a point less than this minimum distance.
  • If the display cannot be updated to keep up with the data required by the zoom operation, then utilizing a mode in which the update information is sent out of the above-indicated order can be utilized. For example, the missing portions of the curve can be provided in an order that allows an approximation to the final display to be generated by the time the pinch zoom gesture is completed. The approximation is then refined as more data is received. For example, an approximation to the final curve could be generated by sending every nth point to mobile device 10 in the region in which data is missing. These points would then be fitted to a curve that is displayed on display screen 12. After these points have been sent, points that are intermediate to these would then be sent. As additional points arrive to fill in the missing points, the data points are then refitted to a curve and displayed. In this manner, the display appears at its final resolution and updates provide a more detailed view of the data. Since the human eye has difficulty perceiving rapidly changing scenes at high resolution, this form of update hides the underlying “noise” resulting from the approximation of the curve by the incomplete data set.
  • Similarly, during a zoom in operation, the missing data is the data between the existing points on the portion of the display that remains within the region displayed on display screen 12. Refer now to FIGS. 4-6, which illustrate a zoom in operation in which the missing data points are sent in an order that allows an approximation to the final display to be generated before the entire data set needed for the display is sent from the data processor. Referring to FIG. 4, user 14 indicates that the user wishes to zoom in on the portion of curve 53 shown at 54 by executing a zoom in pinch gesture. The final display will include some of the data points that are currently stored in mobile device 10 together with a set of additional points that provide information about curve 53 between the points that are currently being displayed. When the zoom in operation commences, mobile device 10 signals data processor 11 to commence sending data points between the existing data points. By the time the zoom in gesture is completed, mobile device 10 has accumulated the data points in FIG. 5. Exemplary data points are labeled at 41-45. These data points are used to generate an approximation 49 to the final display. As additional data points arrive for locations between the existing data point locations such as data point 46 shown in FIG. 6, display screen 12 is updated to provide a better approximation to the final curve as shown at 51.
  • The above-described embodiments of the present invention have utilized specific touch-enabled gestures operating on a mobile device such as a tablet or smart phone. However, the present invention may be utilized in any data processing system that responds to certain gestures on a touch-enabled display. For the purposes of this discussion, a vector touch gesture is defined to be a gesture that is characterized by a direction and a magnitude in which the exact magnitude is not known until the gesture is completed. The direction, in contrast, can be determined by the data processing system prior to the time at which the magnitude can be ascertained.
  • Refer now to FIG. 7, which is a flow chart for one embodiment of a method according to the present invention. The data processing system that operates the touch-enabled display detects a touch gesture that is initiated on the touch-enabled display and determines if the gesture is a vector gesture as shown at 71. If the touch gesture is a vector touch gesture, the data processing system determines the direction associated with that touch direction as shown at 72. In addition, the data processing system determines a set of vector touch gestures that could be the one being implemented as shown at 73.
  • It should be noted that the specific vector gesture may not be exactly known until the vector gesture is completed. For example, a scrolling operation could be performed by a fling or a conventional scroll in which the user moves the user's finger in a continuous manner while in contact with the screen but stops the motion before lifting the finger. A fling, in contrast, is characterized by the finger being in motion at a speed greater than some threshold speed when the finger is lifted. The data processing system cannot determine which scrolling operation is being requested until the user actually lifts the user's finger.
  • After the data processing system determines the set of possible gestures that could be in progress, the data processing system uses the determined direction and the current state of the gesture to execute a command that will at least partially perform work needed to respond to one or more of the possible gestures as shown at 74. For example, consider the case in which all of the possible gestures require a display to be scrolled from its current state to a final state, the scrolling being performed in the determined direction. When the direction is determined, the finger position or positions will have advanced to a new location that provides some additional information about the magnitude associated with the vector gesture. For example, from the current position, the data processing system can determine the minimum distance that the display will need to be scrolled, and hence, the minimum amount of new information needed to execute the scroll. A maximum distance the display could be scrolled in the current possible gesture can also be determined. For example, if the current gesture were a conventional scroll, the maximum distance is determined by measuring the scroll distance corresponding to the user's finger reaching the edge of the screen in the current direction. If the scroll operation was a fling, the maximum distance would be determined by the current scroll speed and some estimate for the maximum speed of the finger when it is lifted.
  • In the case of a scroll operation, the command returns data needed to provide the display in a scrolled state that is a better approximation to the final state than the current state. In general, the system is updated based on the results of those commands as shown at 75. For example, an updated display could be generated based on the results of the executed commands as the data from those commands is received. The commands can continue to elicit data needed to execute the maximum possible scroll in the case of a scroll. If the scroll requires less than this amount of data, the excess data can be ignored or saved to facilitate a future operation.
  • After the commands are executed, the data processing system determines if the gesture has been completed as shown at 76. If the gesture has not been completed, the current state of the gesture is determined as shown at 77. The state is determined from the current locations of the user's fingers. This information is used to update the possible set of gestures and determine further commands to be executed based on the current estimate of the fingers.
  • If the data processing system determines that the gesture is finished, the data processing completes the operations specified by the gesture as shown at 78. The completion may require that the results from one or more commands that were previously issued be received. In addition, the completion may require additional commands. For example, in the case of a fling gesture, additional data may need to be requested before the screen can be updated to its final form.
  • The above-described examples are directed to actions that result in an updated display. However, actions that control a device that is in communication with the data processing system could also be utilized. In this case, the command could be a command that is sent to an external device that causes that device to carry out an operation. The operation in question could result in an updated display; however, such an update is not necessarily required. For example, an instrument that is separate from the mobile device may be updated without changing the display on the mobile device.
  • The above-described embodiments have utilized a mobile device with a touch-enabled screen. However, the present invention can be practiced on any data processing system that has a touch-enabled screen and software to detect one or more vector gestures. Touch-enabled screens are available on desktop systems as well as mobile devices. It should also be noted that the screen view being updated need not be the touch-enabled screen. A system with two screens, one touch enabled and the other a conventional screen could benefit from the present invention being utilized to receive the vector gestures on the touch-enabled screen, while updating the conventional display.
  • The present invention also includes any computer readable medium that stores instructions that cause a computer reading that medium to execute the method of the present invention. For the purposes of the present discussion, any computer readable medium that qualifies as patentable subject matter under 35 U.S.C. 101 will be defined to be a computer readable storage medium. Such media include non-transitory storage media such as non-volatile memories, magnetic and optically-based disk drives and storage elements read by those disk drives.
  • The above-described embodiments of the present invention have been provided to illustrate various aspects of the invention. However, it is to be understood that different aspects of the present invention that are shown in different specific embodiments can be combined to provide other embodiments of the present invention. In addition, various modifications to the present invention will become apparent from the foregoing description and accompanying drawings. Accordingly, the present invention is to be limited solely by the scope of the following claims.

Claims (20)

What is claimed is:
1. A method of operating a first data processing system having a screen for responding to a vector gesture performed on that screen, said method comprising:
estimating a direction and assuming an identity for said vector gesture prior to said vector gesture being completed;
executing a command that depends on said estimated direction and assumed identity, said command being required to update a current state of said first data processing system if said assumed identity of said vector gesture is correct, said command being executed prior to said vector gesture being completed; and
updating said current state of said first data processing system based on a result generated by said command.
2. The method of claim 1 wherein said screen comprises a touch-enabled screen.
3. The method of claim 1 further comprising:
determining if said vector gesture has been completed; and
updating said first data processing system to a state indicated by said completed vector gesture if said vector gesture was completed.
4. The method of claim 1 wherein said command causes a second data processing system to provide data to said first data processing system.
5. The method of claim 1 wherein updating said current state of said first data processing system comprising updating a display on said first data processing system with data obtained from executing said command.
6. The method of claim 1 wherein said current state comprises a first scene on said screen and wherein said assumed identity comprising a scrolling gesture that indicates that said first scene is to be scrolled in said estimated direction.
7. The method of claim 6 wherein said current state comprises displaying a partially scrolled scene on said screen, said update to said current state comprises a partially scrolled version of said first scene.
8. The method of claim 3 wherein updating said first data processing system to a state indicated by said completed vector gesture comprises displaying a first scene on said screen followed by a second scene having a higher resolution than said first scene.
9. The method of claim 3 wherein estimating a direction and assuming an identity for said vector gesture is repeated if said vector gesture was not completed.
10. The method of claim 1 wherein said first data processor determines an end point for a display to be provided on said screen in response to a fling gesture, and wherein said first data processing system displays intermediate displays that are not part of a final display that would be provided if said fling gesture is not terminated prematurely in a low bandwidth form to allow said fling to be terminated prematurely.
11. A computer readable medium that includes instructions that cause a data processing system to execute a method for responding to a vector gesture performed on a screen on that data processing system, said method comprising:
estimating a direction and assuming an identity for said vector gesture prior to said vector gesture being completed;
executing a command that depends on said estimated direction and assumed identity, said command being required to update a current state of said first data processing system if said assumed identity of said vector gesture is correct, said command being executed prior to said vector gesture being completed; and
updating said current state of said first data processing system based on a result generated by said command.
12. The computer readable medium of claim 11 wherein said screen comprises a touch-enabled screen.
13. The computer readable medium of claim 11 further comprising:
determining if said vector gesture has been completed; and
updating said first data processing system to a state indicated by said completed vector gesture if said vector gesture was completed.
14. The computer readable medium of claim 11 wherein said command causes a second data processing system to provide data to said first data processing system.
15. The computer readable medium of claim 11 wherein updating said current state of said first data processing system comprises updating a display on said first data processing system with data obtained from executing said command.
16. The computer readable medium of claim 11 wherein said current state comprises a first scene on said screen and wherein said assumed identity comprising a scrolling gesture that indicates that said first scene is to be scrolled in said estimated direction.
17. The computer readable medium of claim 16 wherein said current state comprises displaying a partially scrolled scene on said screen, said update to said current state comprises a partially scrolled version of said first scene.
18. The computer readable medium of claim 13 wherein updating said first data processing system to a state indicated by said completed vector gesture comprises displaying a first scene on said screen followed by a second scene having a higher resolution than said first scene.
19. The computer readable medium of claim 13 wherein estimating a direction and assuming an identity for said vector gesture is repeated if said vector gesture was not completed.
20. The computer readable medium of claim 11 wherein said first data processing system determines an end point for a display to be provided on said screen in response to a fling gesture, and wherein said first data processor displays intermediate displays that are not part of a final display that would be provided if said fling gesture is not terminated prematurely in a low bandwidth form to allow said fling to be terminated prematurely.
US13/750,536 2013-01-25 2013-01-25 Method for Utilizing Projected Gesture Completion to Improve Instrument Performance Abandoned US20140215382A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US13/750,536 US20140215382A1 (en) 2013-01-25 2013-01-25 Method for Utilizing Projected Gesture Completion to Improve Instrument Performance
CN201310724114.6A CN103970439A (en) 2013-01-25 2013-12-25 Method for utilizing projected gesture completion to improve instrument performance
CN202210128082.2A CN114779921A (en) 2013-01-25 2013-12-25 Method for improving instrument performance by using completion of predicted gestures
US17/035,610 US11379114B2 (en) 2013-01-25 2020-09-28 Method for utilizing projected gesture completion to improve instrument performance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/750,536 US20140215382A1 (en) 2013-01-25 2013-01-25 Method for Utilizing Projected Gesture Completion to Improve Instrument Performance

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/035,610 Continuation US11379114B2 (en) 2013-01-25 2020-09-28 Method for utilizing projected gesture completion to improve instrument performance

Publications (1)

Publication Number Publication Date
US20140215382A1 true US20140215382A1 (en) 2014-07-31

Family

ID=51224461

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/750,536 Abandoned US20140215382A1 (en) 2013-01-25 2013-01-25 Method for Utilizing Projected Gesture Completion to Improve Instrument Performance
US17/035,610 Active US11379114B2 (en) 2013-01-25 2020-09-28 Method for utilizing projected gesture completion to improve instrument performance

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/035,610 Active US11379114B2 (en) 2013-01-25 2020-09-28 Method for utilizing projected gesture completion to improve instrument performance

Country Status (2)

Country Link
US (2) US20140215382A1 (en)
CN (2) CN114779921A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150128087A1 (en) * 2013-11-01 2015-05-07 Samsung Electronics Co., Ltd. Method and apparatus for displaying page in terminal
US20150350143A1 (en) * 2014-06-01 2015-12-03 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US9898162B2 (en) 2014-05-30 2018-02-20 Apple Inc. Swiping functions for messaging applications
EP3531259A1 (en) 2018-02-23 2019-08-28 Rohde & Schwarz GmbH & Co. KG Measurement device and method for controlling the same
US10620812B2 (en) 2016-06-10 2020-04-14 Apple Inc. Device, method, and graphical user interface for managing electronic communications
US10628011B2 (en) * 2013-12-31 2020-04-21 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Medical devices, method and apparatus for adjusting a time range of a trend chart
US20210357034A1 (en) * 2020-05-15 2021-11-18 Thu Ha TRINH Method and system for processing gestures detected on a display screen of a foldable device
US11188168B2 (en) 2010-06-04 2021-11-30 Apple Inc. Device, method, and graphical user interface for navigating through a user interface using a dynamic object selection indicator
US11740622B2 (en) * 2019-06-12 2023-08-29 Ford Global Technologies, Llc Remote trailer maneuver-assist

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030122787A1 (en) * 2001-12-28 2003-07-03 Philips Electronics North America Corporation Touch-screen image scrolling system and method
US20050206658A1 (en) * 2004-03-18 2005-09-22 Joshua Fagans Manipulation of image content using various image representations
US20090100383A1 (en) * 2007-10-16 2009-04-16 Microsoft Corporation Predictive gesturing in graphical user interface
US20120284664A1 (en) * 2011-05-06 2012-11-08 Microsoft Corporation Caching intermediate data for scroll view rendering
US20130222329A1 (en) * 2012-02-29 2013-08-29 Lars-Johan Olof LARSBY Graphical user interface interaction on a touch-sensitive device
US9052819B2 (en) * 2012-01-25 2015-06-09 Honeywell International Inc. Intelligent gesture-based user's instantaneous interaction and task requirements recognition system and method

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6920619B1 (en) * 1997-08-28 2005-07-19 Slavoljub Milekic User interface for removing an object from a display
US7426696B1 (en) * 2005-07-15 2008-09-16 Minerva Networks, Inc. Method and system for image windowing
KR101304461B1 (en) * 2006-12-04 2013-09-04 삼성전자주식회사 Method and apparatus of gesture-based user interface
US8681104B2 (en) * 2007-06-13 2014-03-25 Apple Inc. Pinch-throw and translation gestures
JP2009140119A (en) 2007-12-05 2009-06-25 Sharp Corp Graphic display device and graphic display method
US20090300554A1 (en) * 2008-06-03 2009-12-03 Nokia Corporation Gesture Recognition for Display Zoom Feature
CN102257471B (en) * 2008-10-26 2015-07-22 思杰***有限公司 Method for rendering a window on a native display of a mobile computing device
US20100177051A1 (en) * 2009-01-14 2010-07-15 Microsoft Corporation Touch display rubber-band gesture
KR101567785B1 (en) 2009-05-28 2015-11-11 삼성전자주식회사 Apparatus and method for controlling zoom function of a portable terminal
US8271898B1 (en) * 2009-06-04 2012-09-18 Mellmo Inc. Predictive scrolling
JP2011170834A (en) * 2010-01-19 2011-09-01 Sony Corp Information processing apparatus, operation prediction method, and operation prediction program
WO2011152001A1 (en) * 2010-06-03 2011-12-08 パナソニック株式会社 Scrolling device, scrolling method, scrolling program and integrated circuit
CN103380405A (en) * 2010-12-30 2013-10-30 汤姆逊许可公司 User interface, apparatus and method for gesture recognition
CN102508573B (en) * 2011-10-28 2015-04-22 Tcl通讯科技(成都)有限公司 Method and equipment for realizing rapid positioning through swiping

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030122787A1 (en) * 2001-12-28 2003-07-03 Philips Electronics North America Corporation Touch-screen image scrolling system and method
US20050206658A1 (en) * 2004-03-18 2005-09-22 Joshua Fagans Manipulation of image content using various image representations
US20090100383A1 (en) * 2007-10-16 2009-04-16 Microsoft Corporation Predictive gesturing in graphical user interface
US20120284664A1 (en) * 2011-05-06 2012-11-08 Microsoft Corporation Caching intermediate data for scroll view rendering
US9052819B2 (en) * 2012-01-25 2015-06-09 Honeywell International Inc. Intelligent gesture-based user's instantaneous interaction and task requirements recognition system and method
US20130222329A1 (en) * 2012-02-29 2013-08-29 Lars-Johan Olof LARSBY Graphical user interface interaction on a touch-sensitive device

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11709560B2 (en) 2010-06-04 2023-07-25 Apple Inc. Device, method, and graphical user interface for navigating through a user interface using a dynamic object selection indicator
US11188168B2 (en) 2010-06-04 2021-11-30 Apple Inc. Device, method, and graphical user interface for navigating through a user interface using a dynamic object selection indicator
US20150128087A1 (en) * 2013-11-01 2015-05-07 Samsung Electronics Co., Ltd. Method and apparatus for displaying page in terminal
US9851886B2 (en) * 2013-11-01 2017-12-26 Samsung Electronics Co., Ltd Method and apparatus for displaying page in terminal
US10628011B2 (en) * 2013-12-31 2020-04-21 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Medical devices, method and apparatus for adjusting a time range of a trend chart
US10739947B2 (en) 2014-05-30 2020-08-11 Apple Inc. Swiping functions for messaging applications
US9898162B2 (en) 2014-05-30 2018-02-20 Apple Inc. Swiping functions for messaging applications
US11226724B2 (en) 2014-05-30 2022-01-18 Apple Inc. Swiping functions for messaging applications
US10416882B2 (en) 2014-06-01 2019-09-17 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US9971500B2 (en) * 2014-06-01 2018-05-15 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US11068157B2 (en) 2014-06-01 2021-07-20 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US11868606B2 (en) 2014-06-01 2024-01-09 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US11494072B2 (en) 2014-06-01 2022-11-08 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US20150350143A1 (en) * 2014-06-01 2015-12-03 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US10620812B2 (en) 2016-06-10 2020-04-14 Apple Inc. Device, method, and graphical user interface for managing electronic communications
EP3531259A1 (en) 2018-02-23 2019-08-28 Rohde & Schwarz GmbH & Co. KG Measurement device and method for controlling the same
US11740622B2 (en) * 2019-06-12 2023-08-29 Ford Global Technologies, Llc Remote trailer maneuver-assist
US20210357034A1 (en) * 2020-05-15 2021-11-18 Thu Ha TRINH Method and system for processing gestures detected on a display screen of a foldable device
US11481035B2 (en) * 2020-05-15 2022-10-25 Huawei Technologies Co., Ltd. Method and system for processing gestures detected on a display screen of a foldable device

Also Published As

Publication number Publication date
US11379114B2 (en) 2022-07-05
US20210011619A1 (en) 2021-01-14
CN114779921A (en) 2022-07-22
CN103970439A (en) 2014-08-06

Similar Documents

Publication Publication Date Title
US11379114B2 (en) Method for utilizing projected gesture completion to improve instrument performance
EP3279763B1 (en) Method for controlling display and electronic device
US20230324196A1 (en) Device, Method, and Graphical User Interface for Synchronizing Two or More Displays
US11698721B2 (en) Managing an immersive interface in a multi-application immersive environment
US8847978B2 (en) Information processing apparatus, information processing method, and information processing program
US9898180B2 (en) Flexible touch-based scrolling
KR102061360B1 (en) User interface indirect interaction
KR101395426B1 (en) Method, apparatus and computer program product for creating graphical objects with desired physical features for usage in animations
WO2018120084A1 (en) Flexible display device control method and apparatus
EP2442220A1 (en) Touch-enebled circle control for time and date entry
KR101919009B1 (en) Method for controlling using eye action and device thereof
US20140282154A1 (en) Method for processing a compound gesture, and associated device and user terminal
CN113824998A (en) Music user interface
US9575578B2 (en) Methods, devices, and computer readable storage device for touchscreen navigation
CN106445357B (en) A kind of slider control control method and mobile terminal
US20150070283A1 (en) Techniques for providing a scrolling carousel
EP3336675A1 (en) Electronic device and input method of electronic device
KR20160101605A (en) Gesture input processing method and electronic device supporting the same
KR20170040706A (en) Device For Providing One-Handed Manipulation User Interface and Method Thereof
KR102346565B1 (en) Multiple stage user interface
CN108920266A (en) program switching method, intelligent terminal and computer readable storage medium
KR101294201B1 (en) Portable device and operating method thereof
WO2013157013A1 (en) Selection of user interface elements of a graphical user interface
US20150253980A1 (en) Information processing method and electronic device
CN106445359A (en) Control controlling method and device

Legal Events

Date Code Title Description
AS Assignment

Owner name: AGILENT TECHNOLOGIES, INC., COLORADO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ENGEL, GLENN R.;REEL/FRAME:029696/0561

Effective date: 20130123

AS Assignment

Owner name: KEYSIGHT TECHNOLOGIES, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AGILENT TECHNOLOGIES, INC.;REEL/FRAME:033746/0714

Effective date: 20140801

STCV Information on status: appeal procedure

Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS

STCV Information on status: appeal procedure

Free format text: BOARD OF APPEALS DECISION RENDERED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION