TW200847001A - Gesturing with a multipoint sensing device - Google Patents

Gesturing with a multipoint sensing device Download PDF

Info

Publication number
TW200847001A
TW200847001A TW097103534A TW97103534A TW200847001A TW 200847001 A TW200847001 A TW 200847001A TW 097103534 A TW097103534 A TW 097103534A TW 97103534 A TW97103534 A TW 97103534A TW 200847001 A TW200847001 A TW 200847001A
Authority
TW
Taiwan
Prior art keywords
action
touch
input
finger
event
Prior art date
Application number
TW097103534A
Other languages
Chinese (zh)
Inventor
Wayne C Westerman
Steven P Hotelling
Myra Haggerty
Bas Ording
Nima Parivar
Duncan Robert Kerr
Original Assignee
Apple Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apple Inc filed Critical Apple Inc
Publication of TW200847001A publication Critical patent/TW200847001A/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/038Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/038Indexing scheme relating to G06F3/038
    • G06F2203/0381Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Position Input By Displaying (AREA)
  • Input From Keyboards Or The Like (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

Methods and systems for implementing gestures with sensing devices are disclosed. More particularly, methods and systems related to gesturing with multipoint sensing devices are disclosed.

Description

200847001 九、發明說明: 【發明所屬之技術領域】 本發明一般係關於與多點感測裝置相關聯的示意動作。 本申請案主張2006年1月30曰申請之美國臨時專利申請 案第 60/763,605 號標題為&quot;GESTURING WITH A MULTIPOINT SENSING DEVICE&quot;之優先權權利,該案以 引用的方式併入本文中。 本申請案是2004年7月30日申請之美國申請案第 10/903,964號標題為&quot;GESTURES FOR TOUCH SENSITIVE INPUT DEVICES”的部份接續申請案,並且該案以引用的 方式併入本文中。 本申請案是2004年12月17曰申請之美國申請案第 1 1/015,434 號標題為’’METHOD AND APPARATUS FOR INTEGRATING MANUAL INPUT”(該案主張2001年 11 月 27 曰頒布之美國專利案第6,323,846號之優先權)的部份接續 申請案,該案主張1998年1月26曰申請之美國臨時專利申 請案第60/072,509號之優先權,彼等案皆以引用的方式併 入本文中。 本申請案亦相關於下列專利申請案,彼等案皆以引用的 方式併入本文中: 2003年9月2曰申請之美國專利申請案第10/654,108號標 題為”AMBIDEXTROUS MOUSE” ; 2004年2月27日申請之美國專利申請案第1〇/789,676號標 題為”SHAPE DETECTING INPUT DEVICE” ; 128630.doc 200847001 2004年5月6日申請之美國專利申請案第1〇/840,862號標 題為”MULTIPOINT TOUCHSCREEN” ; 2005年4月26曰申請之美國專利申請案第1 1/1 15,539號標 題為 ” HAND HELD ELECTRONIC DEVICE WITH MULTIPLE TOUCH SENSING DEVICES11 ;200847001 IX. Description of the Invention: TECHNICAL FIELD OF THE INVENTION The present invention generally relates to illustrative actions associated with multi-point sensing devices. The present application claims the benefit of priority to &quot;GESTURING WITH A MULTIPOINT SENSING DEVICE&quot; in U.S. Provisional Patent Application Serial No. 60/763,605, filed on Jan. 30, 2006, which is incorporated herein by reference. This application is a continuation-in-part of U.S. Application Serial No. 10/903,964, entitled &quot;GESTURES FOR TOUCH SENSITIVE INPUT DEVICES, filed on July 30, 2004, which is incorporated herein by reference. This application is filed on December 17, 2004, in the U.S. Application Serial No. 1 1/015,434 entitled "'METHOD AND APPARATUS FOR INTEGRATING MANUAL INPUT" (the case claims US Patent No. 6,323,846 issued November 27, 2001) </ RTI> </ RTI> <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; The present application is also related to the following patent applications, each of which is incorporated herein by reference: U.S. Patent Application Serial No. 10/654,108, filed on Sep. 2, 2003, entitled "AMBIDEXTROUS MOUSE"; 2004 U.S. Patent Application Serial No. 1/789,676, filed on Feb. 27, entitled "SHAPE DETECTING INPUT DEVICE"; 128 630.doc 200847001 U.S. Patent Application Serial No. 1/840,862, filed on May 6, 2004, is entitled " U.S. Patent Application Serial No. 1 1/1 15,539, entitled "HAND HELD ELECTRONIC DEVICE WITH MULTIPLE TOUCH SENSING DEVICES11;

2004年7月30日申請之美國專利申請案第11/241,839號標 題為,’PROXIMITY DETECTOR IN HANDHELD DEVICE” ; 2005年3月4日申請之美國臨時專利申請案第60/65 8,777 號標題為&quot;MULTI-FUNCTIONAL HAND-HELD DEVICE&quot;; 2004年7月30日申請之美國專利申請案第10/903,964號標 題為&quot;GESTURES FOR TOUCH SENSITIVE INPUT DEVICES,,; 2005年1月18日申請之美國專利申請案第1 1/03 8,590號標 題為&quot;MODE-BASED GRAPHICAL USER INTERFACES FOR TOUCH SENSITIVE INPUT DEVICES&quot;; 2005年1月31曰申請之美國專利申請案第11/048,264號標 題為,’GESTURES FOR TOUCH SENSITIVE INPUT DEVICES” ; 2005年9月16日申請之美國專利申請案第1 1/228,737號標 題為”ACTIVATING VIRTUAL KEYS OF A TOUCH· SCREEN VIRTUAL KEYBOARD” ; 2〇05年9月16曰申請之美國專利申請案第ll/228,758號標 題為&quot;VIRTUAL INPUT DEVICE PLACEMENT ON A TOUCH SCREEN USER INTERFACE” ; 128630.doc 200847001 2005年9月16日申請之美國專利申請案第11/228,700號標 題為&quot;OPERATION OF A COMPUTER WITH TOUCH SCREEN INTERFACE,’ ; 2004年8月26日申請之美國專利申請案第10/927,925號標 題為,,VISUAL EXPANDER” ; 2004年8月25日申請之美國專利申請案第10/927,575號標 題為”WIDE TOUCHPAD ON A PORTABLE COMPUTER&quot;; 2007年1月3日申請之美國專利申請案第1 1/619,553號標 題為&quot;MULTI-TOUCH GESTURE DICTIONARY&quot;;及 2007年1月3日申請之美國專利申請案第11/619,571號標 題為”MULTI-TOUCH GESTURE DICTIONARY&quot;。 【先前技術】 現在存在用於在電腦系統内實行操作的許多樣式之輸入 裝置。操作一般對應於在顯示螢幕上移動游標及作出選 擇。操作亦可包括分頁、捲動、平移、縮放等。例如,輸 入裝置可包括按鈕、開關、鍵盤、滑鼠、執跡球、觸控 板、搖桿、觸控螢幕等等。該等裝置之各個具有在設計電 腦系統時予以考慮的優點及缺點。 關於觸控板,當沿觸控板表面移動手指時,輸入指標之 移動對應於使用者手指(或觸控筆)之相對移動。另一方 面,觸控螢幕係具有覆蓋螢幕之觸敏透明面板的一類型之 顯示螢幕。當使用觸控螢幕時,使用者藉由直接指向(通 常採用觸控筆或手指)螢幕上之GUI物件在顯示螢幕上作出 選擇。一般而言,觸控裝置辨識觸控及觸控位置,而計算 128630.doc 200847001 系統解譯觸控,之後根據觸控事件實行動作。 為提供額外功能性’採用該等輸入裝置之某些實施示意 動作。例如,觸控板中,當在觸控板表面上偵測一或多個 敲擊時作出選擇。某些情形中,可敲擊觸控板之任何部 分,其他情形中,可敲擊觸控板之專用部分。除選擇外, 可藉由在觸控板邊緣使用手指運動起始捲動。 不幸的是,示意動作嚴重受限於該事實:即使當在感測 表面上放置多個物件,大多數觸控技術僅能夠報告單點。 即,其缺少同時追蹤多個接觸點之能力。在電阻及電容技 術中,決定所有同時出現之觸控點的平均值,並報告落於 觸控點之間的某處之單點。在表面波及紅外線技術中,由 於遮蔽,不可能分辨落於相同水平或垂直線上的多個觸控 點之準確位置。任一情形中,產生錯誤結果。 根據上述内容,需要多點感測裝置及實施具有多點感測 裝置之示意動作之方法。 【發明内容】 一項具體實施例中,本發明係關於電子系統。該電子系 統包括一多點感測裝置,其提供用於從一或多個物件接收 輸入之一多點感測區域。該電子系統亦包括一示意動作模 組,其經組悲用以決定用於藉由該多點感測裝置之該多點 感測區域接收之一給定輸入配置的一示意動作集,用以監 視用於包括於該示意動作集内之一或多個示意動作集的該 給定輸入配置,以及用以在採用該輸入配置實行該示意動 作事件時起始與一示意動作事件相關聯之輸入動作。例 128630.doc 200847001 如,該輸入配置可為手指及/或手其他部分之配置。 另-具體實施例中,本發明係關於示意動作控制方法。 該方法包括同時偵測一感測區域内之多個點。該方法亦包 括當在該感測區域内偵測一或多個點時決定一弦(chH)。 該弦係該感測區域内的一特定點配置。該方法進一步包括 決定-示意動作集,其將命令與一或多個示意動作事件相 關聯。該方法額外地包括針對示意動作事件監視點。另 外,該方法包括即使辨識一示意動作事件,實行與示意動 作相關聯之命令。 另-具體實施例中,本發明係關於控制操作。㈣㈣ 包括價測一觸控或接近觸控。操作亦包括決定用於觸控之 不意動作集。示意動作集包括用於引發或起始一命令之一 或多個示意動作事件。操作進一 一 ^ ^ ^ ^ ^ /匕栝針對一不意動作事 一 匕枯田貝仃與不意動作集相關聯 之一示意動作事件時起始一命令。U.S. Patent Application Serial No. 11/241,839, filed on July 30, 2004, entitled &quot;PROXIMITY DETECTOR IN HANDHELD DEVICE&quot;; U.S. Provisional Patent Application No. 60/65,777, filed on March 4, 2005 U.S. Patent Application Serial No. 10/903,964, filed on July 30, 2004, entitled &quot;GESTURES FOR TOUCH SENSITIVE INPUT DEVICES,,; January 18, 2005 U.S. Patent Application Serial No. 1 1/03, 590, entitled &quot;MODE-BASED GRAPHICAL USER INTERFACES FOR TOUCH SENSITIVE INPUT DEVICES&quot;; US Patent Application Serial No. 11/048,264, filed January 31, 2005, entitled &quot;GESTURES FOR TOUCH SENSITIVE INPUT DEVICES"; US Patent Application No. 1 1/228,737, filed on September 16, 2005, entitled "ACTIVATING VIRTUAL KEYS OF A TOUCH· SCREEN VIRTUAL KEYBOARD"; September 16, 2005 US Patent Application No. ll/228,758 is entitled &quot;VIRTUAL INPUT DEVICE PLACEMENT ON A TOUCH SCREEN USER INTERFACE"; 128630.doc 200847001 September 16, 2005 U.S. Patent Application Serial No. 11/228,700, entitled &quot;OPERATION OF A COMPUTER WITH TOUCH SCREEN INTERFACE,'; U.S. Patent Application Serial No. 10/927,925, filed on August 26, 2004, entitled, VISUAL EXPANDER U.S. Patent Application Serial No. 10/927,575, filed on Aug. 25, 2004, entitled &quot;WIDE TOUCHPAD ON A PORTABLE COMPUTER&quot;; U.S. Patent Application Serial No. 1/619,553, filed on Jan. 3, 2007, entitled &quot;;MULTI-TOUCH GESTURE DICTIONARY&quot;; and U.S. Patent Application Serial No. 11/619,571, filed on Jan. 3, 2007, entitled &quot;MULTI-TOUCH GESTURE DICTIONARY&quot;. [Prior Art] There are now many styles of input devices for performing operations within a computer system. The operation generally corresponds to moving the cursor on the display screen and making a selection. Operations can also include paging, scrolling, panning, zooming, and the like. For example, the input device can include a button, a switch, a keyboard, a mouse, a walking ball, a touch pad, a joystick, a touch screen, and the like. Each of these devices has advantages and disadvantages that are considered when designing a computer system. With regard to the touchpad, when the finger is moved along the surface of the touchpad, the movement of the input index corresponds to the relative movement of the user's finger (or stylus). On the other hand, the touch screen has a type of display screen that covers the touch sensitive transparent panel of the screen. When using a touch screen, the user makes a selection on the display screen by pointing directly to the GUI object on the screen (usually using a stylus or finger). In general, the touch device recognizes the touch and touch position, and calculates the 128630.doc 200847001 system to interpret the touch, and then performs the action according to the touch event. To provide additional functionality, some of the implementations of the input devices are used to illustrate the actions. For example, in a touchpad, a selection is made when one or more taps are detected on the surface of the trackpad. In some cases, you can tap any part of the touchpad, and in other cases, tap the dedicated portion of the touchpad. In addition to selection, scrolling can be initiated by using finger motion at the edge of the trackpad. Unfortunately, the gesture is severely limited by the fact that most touch technologies are only capable of reporting a single point even when multiple objects are placed on the sensing surface. That is, it lacks the ability to track multiple touch points simultaneously. In resistance and capacitance techniques, determine the average of all simultaneous touch points and report a single point that falls somewhere between the touch points. In surface wave and infrared technology, it is impossible to distinguish the exact positions of multiple touch points falling on the same horizontal or vertical line due to shadowing. In either case, an erroneous result is produced. In view of the above, a multi-point sensing device and a method of performing a schematic action with a multi-point sensing device are required. SUMMARY OF THE INVENTION In one embodiment, the present invention is directed to an electronic system. The electronic system includes a multi-point sensing device that provides a multi-point sensing region for receiving input from one or more objects. The electronic system also includes a schematic action module for determining a set of gestures for receiving a given input configuration by the multi-point sensing region of the multi-point sensing device. Monitoring the given input configuration for one or more sets of gestures included in the set of gestures, and for initiating an input associated with a gesture event when the gesture action is performed using the input configuration action. Example 128630.doc 200847001 For example, the input configuration can be configured for the finger and/or other parts of the hand. In another embodiment, the invention is directed to a schematic motion control method. The method includes simultaneously detecting a plurality of points within a sensing region. The method also includes determining a chord (chH) when one or more points are detected within the sensing region. The chord is a specific point configuration within the sensing region. The method further includes a decision-schematic set of actions that associates the command with one or more gesture events. The method additionally includes monitoring points for illustrative action events. Additionally, the method includes executing a command associated with the gesture action even if a gesture event is identified. In another embodiment, the invention is directed to control operations. (4) (4) Including price measurement, touch or proximity touch. The operation also includes determining the set of unintentional actions for touch. The set of gestures includes one or more gesture events for initiating or initiating a command. Operation into a ^ ^ ^ ^ ^ / 匕栝 for an unintentional action A 匕 田 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃 仃

另一具體實施例中,本發明#關 立 七月係關於不意動作操作。操作 包括監視一觸控運動。操作亦 八#紘¥ π 已括在弟一與弟二狀態間區 刀該觸控運動。操作進一步 II於目I丨每一 &gt; 匕括右觸控運動與第一狀態相In another embodiment, the present invention is related to the July operation regarding unintended operation. Operation includes monitoring a touch motion. Operation also 八#纮¥ π has been included in the division between the younger brother and the younger brother. Operation further II, I, each &gt; 右, the right touch motion and the first state

關果卩則實行第一動作。握你链L 示料#作額外地包括若動 相關聯則實行第二動作。 〜弟一狀慇 另 具體實施例中,本於明旅關狄&amp; 包括裎徂# ^ X ,、關於控制操作。控制操作 匕括棱供弟一輸入裝置及 梦晉—▲ 个丨』於弟一輸入裝置之第二輸入 裝置。弟一輸入裝置包括物件残 入事件之總μ ”壯 例如用於提供輸 f什之觸控感測裝置。瘂从 ’、’、匕括針對輸入事件監視第 128630.doc 200847001 一輸入裝置。操作進一步包括針對輸入事件同時監 輸入裝置。操作額外地包括根據與第一輸入裝置相關聯: 輸入事件實行輸入操作。另外’該方法包括根據與該第二 輸入裝置相關聯的輸入事件同時實行輸入操作。 β另一具體實施例中,本發明係關於控制操作。控制操作 提供一輸入功能清單。輸入功能具有命令及連結至命令之 示意動作事件。命令係關於輸入功能。操作亦包括將輸入 功能指派給弦。操作額外地包括㈣職時將輸人功能連 結至弦。 另一具體實施例中,本發明係關於控制面板。控制面板 包括顯示—輸人功能選單。控制面板亦包括顯示—示意動 作集,其顯示與-較輸人㈣相關聯之命令,以及指派 給該等命令之示意動作事件。控制操作進—步包括顯示一 =早1建置-示意動作映射’―使用者從輸入功能清 早、擇期望輸人功能,並從該弦清單選擇-期望弦。 另一具體實施例中,本發明係關於示意動作設計程序。 =程序包括提供-弦清單並從最簡單至最難分級弦。設 口十程序亦包括從最頻繁$县 I至取不頻繁分級操作頻率。設計程 序進一步包括將最簡單 早弦與取頻繁刼作匹配,並將最難弦 頻繁操作匹配。設計程序可額外地包括精細調諧弦/ 3眾作为組。 動作;r :體只%例中’本發明係關於示意動作操作。示意 扑狀Γ &amp;括制第—手指。示意動作操作亦包括決定手 。例如’手指狀態可為移動或固定。示意動作操作 128630.doc -11 - 200847001 7-步包括偵測一或多個額外手指。例如,可偵测第二手 指。不意動作操作額外地包括決錢外手指之狀態。例 如,額外手指之狀態可為其存在或不存在。另外,該方法 =括根據弟一及額外手指相對於彼此之狀態時序實施不同 =入模式。例如’不同模式可為指向模式、拖$模式等 【實施方式】 作==揭示示意動作及用於實施具有感測裝置之示意動 、、列-置之-:特…,揭示示意動作及實施具有多點感 不思動作的方法。多點感測裝置具有優於傳统單 點裝置之芒早很机 日朴 1寻、、死早 干優』’即其可同時或幾乎同時區別多個物件 /曰)°大多數情形中,制此類裝置之多點感測裝置及 糸統針對觸控或接近觸控事件監視—表面。當此—事件發 生時’其可決定接觸之不同區域並經由其幾; :置識料件性質。-旦被識別,監視觸控或接近觸二 以决疋其是否對應於各種示意動作事件。 可將不意動作事件定義為與映射至一或多個特定 作之感測表面的格式化互動。可透 〜 宣从甚 士裡于疋成不意動作 事件,更特定言之係手指、運動、敲擊、 $ 或類似物。由於表面係基於多 :奋及’ 手私或手之其他接觸部分實行複雜示意動作。事奋 :此原因’可發展出類似於符號語言的較大示二作: 二:如,示意動作語言(或映射)可包括—組指令,复: 識接觸配置(例如弦),辨識示意動作事件(例如運動心 128630.doc •12- 200847001 f通知示意動作事件之—或多個軟體代理者及/或回 不思動作事件而採用何種動作。例如,多點示意動作之 範例可在美國專利申請案第膽3,964、剛8,別、及 1 1/015,434號中找到,其全部以提及方式併入本文。 廣泛的不同示意動作 ^ ^ 。莉邗了與夕點感測裝置一起利用。例 如’不忍動作可為單點或吝 動作m、 思動作;靜態或動態示意 、·^力段不意動作;及/或類似物 作係採用單一接觸赴者一 早2不思動Guan Guojun carries out the first action. Hold your chain L. The material # additionally includes the second action. ~ Brother Yi Yin In another specific example, Ben Ming Guan Di &amp; Include # ^ X, about control operations. Control operation 匕 棱 供 供 供 供 输入 输入 输入 输入 输入 输入 输入 输入 输入 输入 输入 ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ ▲ The input device includes the total μ of the object remnant event. For example, it is used to provide a touch sensing device for the input. 痖 From ', ', for the input event monitoring, 128630.doc 200847001 an input device. Further comprising simultaneously monitoring the input device for the input event. The operation additionally includes associating with the first input device: the input event performs an input operation. Additionally the method includes simultaneously performing an input operation based on an input event associated with the second input device In another specific embodiment, the present invention relates to a control operation. The control operation provides an input function list. The input function has a command and a schematic action event linked to the command. The command is related to the input function. The operation also includes assigning the input function. The operation additionally includes (4) the time to connect the input function to the string. In another embodiment, the invention relates to a control panel. The control panel includes a display-input function menu. The control panel also includes a display-schematic action. Set, which displays the commands associated with the inferior (four) and assigned to them The command action event of the command. The control operation step includes displaying a = early 1 build - indicating the action map ' - the user selects the input function early, selects the desired input function, and selects the desired string from the string list. In a specific embodiment, the present invention relates to a schematic action design procedure. = The program includes providing a list of strings and from the simplest to the most difficult to classify the strings. The program ten program also includes the frequency from the most frequent $ County I to the infrequently graded operation frequency. The design procedure further includes matching the simplest early string with the frequent hits and matching the most difficult strings frequently. The design program can additionally include a fine tuning string/3 group as a group. Action; r: body only in the example The present invention relates to a schematic action operation. The gesture is shown to include the first finger. The gesture action also includes determining the hand. For example, the finger state can be moved or fixed. The gesture operation 128630.doc -11 - 200847001 7 The step includes detecting one or more additional fingers. For example, the second finger can be detected. The unintentional action operation additionally includes the state of the finger outside the money. For example, the state of the extra finger In addition, the method includes implementing different input mode according to the state sequence of the first one and the extra finger relative to each other. For example, the 'different mode may be pointing mode, dragging $ mode, etc. == reveals the gestures and is used to implement the gestures with the sensing device, the column-position-: special... to reveal the schematic actions and implement the method with multiple points of inactivity. The multi-point sensing device is superior to The traditional single-point device has long been a machine, and it can be used to distinguish multiple objects at the same time or almost simultaneously. In most cases, multi-point sensing of such devices is made. The device and the system monitor for touch or proximity touch events - the surface. When this occurs, the event can determine the different areas of contact and pass through it; : Identify the nature of the material. Once identified, the touch or proximity touch is monitored to determine if it corresponds to various gesture events. Unintentional action events can be defined as formatted interactions with one or more specific sensing surfaces. It can be transparent. It is said that it is a distraction action event, more specifically a finger, movement, tapping, $ or the like. Because the surface is based on a lot of: the end of the 'hands or other parts of the hand to carry out complex gestures. The excitement: this reason 'can develop a larger display similar to the symbol language: 2: For example, the gesture language (or mapping) can include - group instructions, complex: contact configuration (such as strings), identify gestures Events (such as Sports Heart 128630.doc • 12- 200847001 f notifications indicate action events - or actions taken by multiple software agents and/or back to action events. For example, examples of multi-point gestures can be found in the US The patent application is found in PCT 3,964, just 8, et al., and 1 1/015, 434, all of which are incorporated herein by reference. PCT PCT PCT. For example, 'can't bear the action can be a single point or awkward action m, thinking action; static or dynamic indication, · force segment unintentional action; and / or similar system using a single contact to go early in the morning 2 do not think

|接觸點實仃之該等示意 自(例如)單-手指、手掌或觸控筆之單一觸控實行= 作。多點示意動作係可採用多個點實行示動:動 例如採用來自(例如)多個手指、手指 羊:;動作’ 肇、客徊總手旱手指及觸控 七一工筆及/或其任何組合之多個觸控實行示咅動 作。靜態示意動作可為實、- 弦)之該等示意動作,而動態示音動 作事件(例如 動作事件(例如運動、敲擊等)之該等 』… 動作可為在單一行φ每—^ ^ 。連續示意 壬中貝仃的該等示意動作,而分坪千立 動作可為在不同牛_ ★ / 刀丰又不忍 仕小门步驟或行程之序列中實 作。儘管本文給出數個&amp;如Λ 丁的^亥專不意動 出數個耗例,應明白此係範 可按各種形式具體化多點感測裝置,包括限t 尺寸觸控板、大型延吾手 《於払準 另外,多點感測裝置可你… 鮮觸敏外罩等。 置了位於許多形狀因數 於平板電腦、膝上i ^ I 包括但不限 H:細、桌上型電腦以 置,例如媒體播放器、PDA、蜂巢式電:持式叶鼻裝 用輸入裝置上發現多 D寺寺。亦可在專 感測衣置,例如觸控螢幕監視器、 128630.doc -13- 200847001 、, ·€板、輸入板、滑鼠等等。多點感測裝置之範例 可在關專利中請案第讀4M62、10/927,575號中找到, 其全部以提及方式併入本文。 , 乡考圖1至5 4論述數個具體實施例。然而,熟習技 術人士應容易地明自,本文針對料圖式給出的詳細說明 二;解忒目的,因為本發明擴展至該等受限具體實施例 之外。 、The point of contact point is indicated by a single touch of a single-finger, palm or stylus. The multi-point gesture action can be performed by using a plurality of points: for example, using, for example, a plurality of fingers, a finger sheep: an action '肇, a customer's master dry finger and a touch seven-one pen and/or any of them The combined multiple touch performs the demonstration action. Static gestures can be such gestures as real and chords, while dynamic voice action events (such as action events (such as motion, tapping, etc.)... actions can be in a single line φ per ^^ ^ Continuously indicating the above-mentioned gestures of the 仃 仃 , , , , , , , 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千 千If the 亥 的 ^ 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 专 亥 亥 亥Multi-point sensing device can be used for you... Fresh touch-sensitive cover, etc. It is placed in many form factors on tablets, laptops, but not limited to H: fine, desktop computers, such as media players, PDA, honeycomb type: Multi-D Temple is found on the input device of the holding leaf nose. It can also be used in the special clothing, such as touch screen monitor, 128630.doc -13- 200847001, Input boards, mice, etc. Examples of multi-point sensing devices can be used in patents It is found in the number 4M62, 10/927, 575, which is incorporated herein by reference in its entirety herein. The detailed description given in the drawings is for the purpose of explanation, as the invention extends beyond the limited embodiments.

θ ’’、、示根據本發明之一項具體實施例的說明性示意動 r】操作1〇。操作10可始於步驟12,在此步驟提供一多 I 、j裝置夕點感測裝置能夠同時偵測多個接觸或接近 :例如,多點感測裝置可包括多觸控感測表面,苴 能夠同時感測觸控表面上之多個物件。或者或此外,多點 =裝置可包括用於感測靠近表面但不觸碰表面之物件的 旎力(例如近接感測)。可將多點感測裝置具體化為觸控螢 幕、觸控板、觸控感測手掌塾、觸敏外罩及/或類似物。 V驟12後,#作可繼續至步驟14,在此步驟決定多點感 測裝置是否偵測到一觸控或接近觸控。若未偵測到觸控, 操作可等待。若偵測到觸控,操作可繼續至步驟16,在此 步驟決定錢控㈣聯之弦。弦可為特定㈣或接近接觸 配置,其可被指派給某些輸人功能性。—般而t,當將物 件放置在觸控感測表面附近或之上時,建立接觸面積並且 該等接觸面積形成可識別之圖案。可採用任何物件分類完 、圖*爿⑹其包括機械特徵,如觸控筆及圖章及/或 -或多隻手之部分,例如手指、拇指、手掌、指節等。 128630.doc •14- 200847001 弦可廣泛變化,並可取決於許多因素,包括觸控表面尺 寸、觸控表面是否係觸控螢幕或觸控板等。另外,弦可基 於未知接觸之數目或已知接觸之特定配置。弦可進—步基 於接觸是否彼此靠近、位於中性位置或分散。弦可進一步 基於接觸是否彼此鄰近或偏移。弦可進一步基於其是否來 自左及/或右手。 厌疋弦可包括分析 v两街碉役取搔迎,巧仏w促衣囬θ '', showing an illustrative schematic in accordance with an embodiment of the present invention. Operation 10 may begin at step 12, where a multi-I, j-device sensing device is capable of simultaneously detecting multiple contacts or proximity: for example, the multi-point sensing device may include a multi-touch sensing surface, It can simultaneously sense multiple objects on the touch surface. Alternatively or in addition, the multi-point = device may include a force (e.g., proximity sensing) for sensing an object that is near the surface but does not touch the surface. The multi-point sensing device can be embodied as a touch screen, a touch pad, a touch-sensing palm, a touch-sensitive cover, and/or the like. After V is 12, # continues to step 14, where it is determined whether the multi-point sensing device detects a touch or proximity touch. If no touch is detected, the operation can wait. If touch is detected, the operation can proceed to step 16, where the money control (four) joint string is determined. The strings can be specific (four) or close contact configurations that can be assigned to certain input functionality. Typically, when the object is placed near or over the touch sensing surface, the contact area is established and the contact areas form an identifiable pattern. Any object can be classified, Figure *爿(6), which includes mechanical features such as stylus and stamp and/or - or parts of multiple hands, such as fingers, thumbs, palms, knuckles, etc. 128630.doc •14- 200847001 Strings can vary widely and can depend on many factors, including the size of the touch surface, whether the touch surface is a touch screen or a touchpad. In addition, the strings may be based on the number of unknown contacts or the particular configuration of the known contacts. The strings can be advanced—steps based on whether the contacts are close to each other, in a neutral position, or dispersed. The strings can be further based on whether the contacts are adjacent or offset from each other. The string can be further based on whether it comes from the left and/or right hand. The disgusting string can include analysis v.

之物件建立的影像)及辨識特定接觸配置。更特定言之, 決疋可包括經由各接觸之幾何特徵及接觸之幾何配置分類 或識別接觸,然後參考期望弦之資料庫(例如示意動作映 射)。若接觸配置匹配期望弦,則可假定接觸配置係期望 弦。該等步驟亦可包括過攄。例如,某些接觸實際上可能 ::弦之部分。例如,在觸控手寫板情形中,可忽略藉由 ⑴成之接觸’因為其充當用於手之手指的支樓基礎 ^且如很難在觸控表面上方將手掌保持延長之時間週期)。 觸控螢幕之平板PC情形中,可忽略藉由至少一隹手 之拇指形成的接觸,因為其主要用於保持平板π/ 下表1中顯示單手贫夕^ Μ ^ 其係藉由範例方式^而非㈣白表1並非詳盡清單, 一接觸,並添加至表…顯例如’可將手掌計數為 單組合建立更多Γ 與來自相反手之相同清 合,某些弦可能出於各種原因而不可行組 性、人體工學、直觀性等。 匕使用合易 128630.doc -15- 200847001 -旦已決定弦,操作可繼續至步驟u 或多個示意動作事件相關聯的示意動料、2 動作集可為將動作連結至示意動作事件 集可取決於決定之弦’但亦可取決於其他因素:=作 置開啟應用程式、應用程式狀態或模式、其他觸位 等。本質上,弦選擇輸入通道,各通道具有與直相^徵 不同示意動作集。某些情形中,出於組 =的The image created by the object) and the identification of the specific contact configuration. More specifically, the decision may include classifying or identifying the contact via the geometric features of the contacts and the geometric configuration of the contacts, and then referring to a library of desired strings (e.g., a schematic action map). If the contact configuration matches the desired chord, it can be assumed that the contact configuration is the desired chord. These steps may also include sputum. For example, some touches may actually be part of the ::string. For example, in the case of a touch tablet, the contact by (1) can be ignored because it acts as a base for the finger of the hand ^ and if it is difficult to keep the palm extended over the touch surface for a period of time). In the case of a touch screen tablet PC, the contact formed by at least one thumb of the hand can be ignored, because it is mainly used to keep the tablet π / shown in the following table 1 is a one-handed ^ ^ 其 ^ by way of example ^ instead of (d) white table 1 is not an exhaustive list, a contact, and added to the table ... show for example 'can count the palm as a single combination to create more Γ with the same clear from the opposite hand, some strings may be for various reasons Not feasible group, ergonomics, intuitive, etc.匕Using Yiyi 128630.doc -15- 200847001 - Once the string has been determined, the operation can continue to step u or multiple schematic action events associated with the schematic action, 2 action set can be linked to the action action event set Depending on the decision string' but can also depend on other factors: = open application, application state or mode, other touches, etc. Essentially, the strings select input channels, each channel having a different set of action actions than the direct phase. In some cases, out of group =

:,作通道與特定輸入功能性相關·,例如導覽操::用 編輯操作、檢視操作、格式化操作、I具操作二 2見操作等。該等輸人功能性之各”具有4_聯 中令’其係與示意動作事件聯結。 動作可為狀態命令或操控命令。狀態命令係單—實施A 令。狀態命令之範例包括新增、開啟、關閉、刪除::: 命名、選擇全部、剪下、複製、貼上、復原、取消復心 鮮組/取消群組、斜體、粗體、底線、上-個/下—個、播 放/暫停、靜音等。狀態命令之其他範例包括啟動特定程 式、啟動特定程式模式、啟動網站、開啟工具列、開啟選 單、箭頭指示、退格、空格、輸入、標籤、大寫鎖定、功 月b縮放放大/縮小、分頁、音量等。操控命令係連續 操控選定物件之命令。操控命令之範例包括指向、追縱、 拖曳、捲動、平移、縮放、調整大小、延伸、分頁、音量 等。應明白,該等範例並非限制,可使用其他命令。 如上所述,不意動作事件可為任何格式化實體動作,其 可在觸控表面上或上方實行。例如,示意動作事件之範例 128630.doc -16- 200847001 可包括運動、敲擊、遂力變化、停留等。某些情形中,可 藉由弦實行示意動作事件。其他情形中,可藉由弦之子集 實打不意動作事件。其他情形中,可藉由除初始弦外或與 其分離之新接觸實行示意動作事件。下面將詳細說明 動作事件。 步驟20中,可針對示意動作事件監視觸控,而在步驟22 中,可決定是否辨識示意動作事件。辨識示意動作事件可 包括分析接觸之觸控特徵(與弦及/或新接觸相關聯之接 觸),識別特定圖宰及夂者+立 m + n考^動作集及其期望示意動作 =早二圖案匹配期望示意動作事件,則可假定圖案 ㈣望不讀作事件。例如,觸控特徵可包括第―等 Γ,例如運動、敲擊、壓力變化、停冑,以及第二等級考 里’例如速度(絕對或相對)、方向( 對或相對)、尺寸(絕對或相 万位(、、、巴 y 持、喝k間(絕對或相對)、 職對或相對)、長度(絕對或相對)、及/或類似物。 =具”施财’可將4動作事件設計或選擇 與其他示意動作事件區別,以 串擾或錯誤(例如,通常需要干二辨:…作事件時的 示意動作事件區別)。 動作事件容易地與其他 '敲擊'壓力、停 之滑動、向右、向左、向:=下舉例:沿任何方向 下緩慢線性擦過;向右、 向上及向下快速線性彈過; 轉;順時針或逆時針快速旋轉;兩個曼旋 陶個或兩個以上接觸之收 128630.doc 17 200847001 縮或膨脹(例如擴展及閉人 示意動作事件可進心 間的分離)。若係敲擊, 讀墼也 下舉例:1/2敲擊、完全敲擊、多 _ 人破擊、快速敲擊、緩慢 ^夕 件可進一步兴如炎a 右係壓力,不意動作事 牛幻為輕壓或重壓。若係停留,千立叙n 可進-步舉例為較長不思動作事件 此清軍*北 疋暫停或較短固定暫停。應明白, 此α早亚非限制, ㈡ 件。例如h 式及方位分類示意動作事 τ例如’線性運動示咅叙从Τ咖:, the channel is related to the specific input function. For example, the navigation operation: use the editing operation, the viewing operation, the formatting operation, the I operation 2, and the operation. Each of the input functions has a "4_联中令" which is associated with a gesture action event. The action can be a status command or a manipulation command. The status command is a single order - an implementation A command. Examples of status commands include new, Open, Close, Delete::: Name, Select All, Cut, Copy, Paste, Restore, Ungroup, Ungroup, Italic, Bold, Bottom Line, Up/Down, Play, Pause, mute, etc. Other examples of status commands include launching a specific program, launching a specific program mode, launching a website, opening a toolbar, opening a menu, arrow indication, backspace, space, input, label, caps lock, power month b zoom / Reduction, paging, volume, etc. The manipulation command is a command to continuously manipulate selected objects. Examples of manipulation commands include pointing, chasing, dragging, scrolling, panning, zooming, resizing, stretching, pagination, volume, etc. It should be understood that These examples are not limiting and other commands may be used. As noted above, an unintended action event may be any formatted entity action that may be on or in the touch surface. For example, an example of an action event 128630.doc -16- 200847001 may include motion, tapping, force change, stay, etc. In some cases, a gesture event may be performed by a string. In other cases, In other cases, the action event can be performed by a new contact other than or separate from the initial string. The action event will be described in detail below. In step 20, the action event can be monitored for the action event. Control, and in step 22, it may be determined whether to identify the gesture event. Identifying the gesture event may include analyzing the touch feature of the contact (contact associated with the string and/or new contact), identifying the particular map and the leader + The m + n test ^ action set and its desired schematic action = early two pattern matching desired action event, then the pattern (4) can be assumed to be read as an event. For example, the touch feature may include the first -, for example, exercise, knock Strikes, pressure changes, stoppages, and second-level exams such as speed (absolute or relative), direction (pair or relative), size (absolute or phase 10,000 (,,, y Hold, drink k (absolute or relative), job or relative), length (absolute or relative), and / or similar. = "Scrow" can design or select 4 action events and other action events The difference is in crosstalk or error (for example, it usually requires two distinctions: ... the difference between the gesture action events when the event is made). The action event is easily associated with other 'knocking' pressures, sliding, right, left, and backward: = Example: Slowly linearly rubbing in any direction; fast linearly moving right, up, and down; turning; rotating clockwise or counterclockwise; two or two or more contacts of 128630.doc 17 200847001 Shrinkage or expansion (for example, expansion and closing of people to indicate action events can be separated into the heart). If you are tapping, you can also read the example: 1/2 tap, full tap, more _ person break, fast Knocking, slow ^ Eve pieces can be further embarrassed as a right pressure, do not care about the action of the cow is a light pressure or heavy pressure. If the system stays, the thousands of narrations can be entered as an example of a longer non-thinking event. This Qingjun* North 疋 pause or a shorter fixed timeout. It should be understood that this α is early and Asian non-restricted, (ii) pieces. For example, the h-type and azimuth classifications indicate the action τ, for example, the 'linear motion 咅 咅 Τ Τ Τ Τ

及向下,並可進牛二 向右、向左、向上 W進一步包括右上、左上、右下、及左下 有==Γ為基本或進階的。基本示意動作可為具 動作可為具有多個 心〜 過後敲擊或線性向右二=動作,例如線性向右擦 者具有辦加停、線性向左擦過;或 、有牦加或減小壓力之旋轉cCW。 任何數目之飭你 進ί5自不思動作可包括 |目之動作。可出於使用容 使用基本示意動作。 处理稷雜性原因而 動作事件,操作可繼續至步驟以, 二=作事件相關聯之動作。步㈣可包括參考示 旦了疋位與辨識之示意動作事件相關聯之動作。一 疋立可起始動作。若動作传狀能入八 令,並且每…欠實…t:狀“令,可分別啟動命 按--樣運作)二作事件(例如示意動作事件如 建立個別;^形中’可將示意動作事件設計成 事用者延伸或繼續示意動作 火,咬者i °多個狀態命令。此在遊戲模式中對快速開 或者在編輯模式中復原工作時,或者在劉覽模式中翻 128630.doc -18· 200847001 頁時可很有幫助。例如,若藉由快逮向右線性擦過起始單 -復原命令,可藉由以一連續運動(觸控期間)快速線性擦 過然後緩慢線性擦過起始重複復原命令。此範例中,根據 緩慢線性擦過重複地起始復原命令(例如在緩慢線性捧過 期間每H)〇 ms啟動復原命令)。若動作係操控命令,可在 不意動作事件期間連續產生命令(例如命令與示意動作事 2間的-對-對應例如’若命令係捲動,只要實行示 意動作事件並根㈣意動作事件特 即可實行捲動。 ^方向等) •二刼作可繼績至步驟Μ,在此步驟決定是否已 ^刀換事件。切換事件可指重設操作或起始弦變化之事 門旦種方式實施切換事件。例如,可藉由在預定時 間里内移除所有接觸予以實施(例如從觸控表 :可藉由在觸控期間改變基弦予以實施(例如 =亦可從相反手添加/移除接觸予以實施(例如: : = 同時另—隻手仍在觸控)。亦可藉* ’ ΊΤ卞以實施(例如,甚挺 〜、 接觸在預設定時間量内保掊Ιϋ 疋)。亦可藉由標準鍵盤或滑 、 ’、、 以實施。亦可經由示意動作事輸入或按叙點選予 。邱丨戸爭件予以實施。芒 事件,操作返回步驟i 2。若 刀、 步驟20。 力未出現切換事件,則操作返回 ,Λ 表1 ··單手弦範例 接觸:標準配置(無手掌或忽略手掌) 任何一個手指 128630.doc 200847001 任何兩個手指 任何三個手指 任何四個手指 拇指+任何手指 拇指+任何兩個手指 拇指+任何三個手指 拇指+四個手指 未知接觸:變化配置(無手掌或忽略手掌) 兩個相鄰手指 兩個不相鄰手指 兩個相鄰手指+—個不相鄰手指 拇指+兩個相鄰手指 拇指+兩個不相鄰手指 拇指+兩個相鄰手指+—個不相鄰手指 任何兩個相鄰併攏手指 任何兩個相鄰張開手指 任何三個相鄰併擺手指 任何三個相鄰張開手指 四個相鄰併搬手指 四個相鄰張開手指 拇指+兩個相鄰併攏手指 拇指+兩個相鄰張開手指 拇指+三個相鄰併攏手指 拇指+三個相鄰張開手指 128630.doc -20- 200847001 拇指+四個相鄰併攏手指 拇指+四個相鄰張開手指 已知接觸(無手掌或忽略手掌) 食指 中指 無名指 小指 食指+中指 食指+無名指 食指+小指 中指+無名指 中指+小指 無名指+小指 拇指+食指 拇指+中指 拇指+無名指 拇指+小指 拇指+食指+中指 拇指+食指+無名指 拇指+食指+小指 拇指+中指+無名指 拇指+中指+小指 拇指+無名指+小指 食指+中指+無名指 128630.doc -21 - 200847001 食指+中指+小指 食指+無名指+小指 中指+無名指+小指 拇指+食指+中指+無名指 拇指+食指+中指+小指 拇指+食指+無名指+小指 拇指+中指+無名指+小指 食指+中指+無名指+小指 拇指+食指+中指+無名指+小指 其他(手掌朝下) 于知握成拳或併攏為手掌 食指+剩餘手指握成拳或併攏為手a 食指”指+剩餘手指握成拳或併二 食指+中指+無名指+小指握成拳或 旱 拇指+剩餘手指握成拳或併攏為手/為手掌And down, and can enter the cow two right, left, up W further including upper right, upper left, lower right, and lower left == Γ is basic or advanced. The basic schematic action may be that the action may be a plurality of hearts ~ after a tap or a linear right to the right = action, for example, a linear right to the wiper has a stop and a stop, linear to the left; or, increase or decrease the pressure Rotate cCW. Any number of you can enter the ί5. Basic gestures can be used for use. To deal with the cause of the complexity and the action event, the operation can continue to the step, and the second action is associated with the event. Step (4) may include referencing the action associated with the indicated action event of the identification. A standing movement can start. If the action can enter the eight orders, and each... is not true...t: "order, you can start the life-press-like operation separately" two events (such as signal action events such as establishing individual; ^ shape in the 'can be indicated The action event is designed to extend or continue to indicate the action fire, bite the i ° multiple state commands. This is in the game mode for quick open or restore work in edit mode, or in the Liu mode to turn 128630.doc - 18· 200847001 Pages can be helpful. For example, if you quickly wipe the start order-recovery command by snapping to the right, you can quickly and linearly wipe the initial repeat by a continuous motion (during touch). Restore command. In this example, the restore command is started repeatedly according to the slow linear wipe (for example, every H during the slow linear hold) 。ms start recovery command. If the action system manipulates the command, the command can be continuously generated during the unintentional action event. (For example, the command and the action action 2 - the pair - for example, if the command is scrolling, the scrolling action can be performed as long as the gesture action event is performed and the root (four) intention action event is performed. ^ Direction, etc.) The second action can be followed by a step to determine whether the event has been changed. The switching event can refer to the reset operation of the reset operation or the start of the string change. For example, by scheduling All contacts are removed during the time (for example, from the touch meter: it can be implemented by changing the base string during touch (eg = can also be added/removed from the opposite hand) (eg: : = at the same time - The hand is still in touch). You can also use * ' ΊΤ卞 to implement (for example, very ~, contact within a preset amount of time). You can also use the standard keyboard or slide, ',, It can be implemented. It can also be input by gesture action or by the point. The Qiu Jun dispute is implemented. The mang event, the operation returns to step i 2. If the knife, step 20. The force does not show a switching event, the operation returns. Λ Table 1 · · One-handed string example contact: standard configuration (no palm or ignore palm) Any one finger 128630.doc 200847001 Any two fingers any three fingers any four fingers thumb + any finger thumb + any two finger thumb + Three finger thumb + four fingers unknown contact: change configuration (no palm or ignore palm) two adjacent fingers two non-adjacent fingers two adjacent fingers + one non-adjacent finger thumb + two adjacent fingers Thumb + two non-adjacent finger thumbs + two adjacent fingers + one non-adjacent finger any two adjacent close fingers any two adjacent open fingers any three adjacent and pendulum fingers any three adjacent Open your fingers four adjacent and move your fingers four adjacent open fingers thumb + two adjacent close fingers thumb + two adjacent open fingers thumb + three adjacent close fingers thumb + three adjacent open Finger 128630.doc -20- 200847001 Thumb + four adjacent close fingers thumb + four adjacent open fingers known contact (no palm or neglect of the palm) index finger middle finger ring finger index finger + middle finger index finger + ring finger index finger + little finger middle finger + Ring finger middle finger + little finger ring finger + little finger thumb + index finger thumb + middle finger thumb + ring finger thumb + little finger thumb + index finger + middle finger thumb + index finger + ring finger thumb + index finger + little finger thumb + middle finger + ring finger thumb + middle finger + Finger thumb + ring finger + little finger index finger + middle finger + ring finger 128630.doc -21 - 200847001 index finger + middle finger + little finger index finger + ring finger + little finger middle finger + ring finger + little finger thumb + index finger + middle finger + ring finger thumb + index finger + middle finger + little finger thumb + index finger + ring finger + little finger thumb + middle finger + ring finger + little finger index finger + middle finger + ring finger + little finger thumb + index finger + middle finger + ring finger + little finger other (palm down) Yu Zhi grip into a fist or close together for the palm index finger + remaining finger grip into a fist or Close together for the hand a forefinger" refers to the remaining finger grip into a fist or two index finger + middle finger + ring finger + little finger grip into a fist or dry thumb + remaining fingers grip into a fist or close together for the hand / for the palm

拇指+食指+剩餘手指握成拳或併攏為手掌 =+:指+中指+剩餘手指握成拳或併攏為” ㈣曰+艮指+中指+無名指+小指握成拳或併攏為 拇指+食指+剩餘手指握成拳或併攏為手掌 拇指+食指+中指+剩餘手指握成拳或併攏為手零 拇丸+艮#曰+中指+無名指+小指握成拳或併攏為 其他 手右側 手左側 128630. d〇c •22- 200847001 手背面 手正面(全部) 其他(指節朝下用力按) 手指握成拳或併攏為手掌 食指伸出+剩餘手指握成拳或併攏為手掌 ::伸出+中指伸出+剩餘手指握成拳或併 广伸出+中指伸出+無名指伸出+小指握成 二 拇指+手指握成拳或併攏為手掌 械為手旱 拇“+艮才曰伸出+剩餘手指握成拳或併攏為手掌 拇指+食指伸出+中指 梅指卿出+中指伸出+=:握成拳或併攏為手掌 為手掌 _出+無名才曰伸出+小指握成拳或併攏 圖2顯示根據本發明之— 作50。操作可始 員爲施例的說明性控制操 控。 、^驟52 ’在此步驟偵測觸控或接近觸 步驟52後,操作可繼續至 定-示意動作集。示意動作 控特徵、觸控位置、開啟應 大多數情形巾,以動作集 配置。 步驟54後,操作可繼續至 動作集相_之示意動作事 括-或多個示意動作事件, 可將命令聯結或連結至特定 步驟54,在此步驟針對觸控決 集可取決於許多因素,包括觸 用程式、應用程式模式等等。 至少部分係基於觸碰時之接觸 步驟56,在此步驟針對與示意 件監視觸控。示意動作集可包 其用於引發或起始命令(例如 示意動作事件)。 128630.doc -23- 200847001 步驟56後,操作可繼續至步驟58,在此步驟當實行示音 事件%起始一或多個命令。例如,使用者可滑動特定 乎才曰配置以引發或起始捲動事件。 圖3顯示根據本發明之一項具體實施例的說明性控制操 例如,控制操作6G可對應於圖2中之步㈣。控制 可始於步驟62,在此步驟辨識初始接觸配置。之 =在步驟64’可將初始接觸配置與'組儲存之接觸配置 :。例如,系統可參考示意動作映射,其包括初始接觸 配置之-清單及指派給其之示意動作集。若存在一匹配, 操作可繼續至步驟%,在 . 接觸配置的示意動作集载經指派給辨識之初始 圖4顯示根據本發明之一項具體實施例的說明性控制操 。控制操作7〇可始於步驟72’在此步驟摘測觸控。之 步驟74,可決定觸控位置。之後,在步驟76,可決 =觸控相關聯之接觸配置(例如觸控圖案)。之後,在步 廡田’可決定作用中應用程式。之後,在步驟80,可決定 之當前狀態(例如當前模式)。之後,在步驟 了根據上述_式矣伽、m 备— 一 ’、疋之屬性(步驟74至80)來設定或選 連結上述^動作集。例如’使用決定之屬性,系統可參考 作^射L性之各者至特定示意動作集的已儲存之示意動 使用藉由預設建立儲存之示意動作映射,或者可由 動作映射:關::::::者:使用控制面板來改變與示意 式化示意動作4值,或使用訓練㈣ning)序列以程 128630.doc -24- 200847001 圖5顯示根據本發明之一項具體實施例 作100。柃舍i i品从 1土 k制刼 一觸批 於步驟102’在此步驟決定是否偵測 上。右偵測到觸控,操作可繼續至步驟104, 驟辨識接觸配置。步驟104可包括子步㈣6及⑽。=Thumb + index finger + remaining finger grip into a fist or close together for the palm = +: finger + middle finger + remaining fingers grip into a fist or close together " (four) 曰 + 艮 finger + middle finger + ring finger + little finger grip into a fist or close together for the thumb + index finger + The remaining fingers are clenched into fists or close together for the palm of the thumb + index finger + middle finger + remaining fingers clenched into fists or close together for the hand zero thumb pill + 艮 #曰 + middle finger + ring finger + little finger grip into a fist or close together for the other hand right hand left side 128630. D〇c •22- 200847001 Hand on the back of the hand (all) Others (pressing the knuckles face down) Fingers clenched into fists or close together for the forefinger of the palm of your hand + Remaining fingers clenched into fists or close together for the palm:: Extension + middle finger Extend + remaining finger grip into a fist or wide out + middle finger extended + ring finger extended + little finger grip into two thumbs + finger grip into a fist or close together for the palm hand for the hand dry thumb "+ 艮 曰 曰 out + remaining Finger grip into a fist or close together for the palm of the thumb + index finger extended + middle finger plum finger out of the middle + middle finger extended +=: grip into a fist or close together for the palm of the hand _ out + no name 曰 曰 小 + little finger grip into a fist or close together Figure 2 shows a 50 according to the invention. The operation can be initiated by an illustrative control operation of the example. After the touch or proximity touch step 52 is detected in this step, the operation can continue to the set-illustrated action set. The action control features, touch position, and opening should be configured in most cases. After step 54, the operation may continue to the action set event - or a plurality of schematic action events, the command may be linked or linked to a specific step 54, where the decision for the touch may depend on a number of factors, Includes apps, app mode, and more. At least in part based on the contact step 56 when touched, the step is to monitor the touch with the gesture. A set of gestures can be used to raise or initiate a command (e.g., to indicate an action event). 128630.doc -23- 200847001 After step 56, operation can continue to step 58 where one or more commands are initiated when the voice event % is implemented. For example, the user can slide the specific configuration to initiate or initiate a scrolling event. Figure 3 shows an illustrative control operation in accordance with an embodiment of the present invention. For example, control operation 6G may correspond to step (4) of Figure 2. Control may begin at step 62 where the initial contact configuration is identified. = In step 64', the initial contact configuration can be configured with the 'group storage' contact: For example, the system may reference a schematic action map that includes a list of initial contact configurations and a set of gestures assigned to it. If there is a match, the operation can continue to step %, where the schematic action set of the contact configuration is assigned to the initial identification. Figure 4 shows an illustrative control operation in accordance with an embodiment of the present invention. Control operation 7 can begin at step 72' where the touch is extracted. In step 74, the touch position can be determined. Thereafter, at step 76, the contact configuration (eg, touch pattern) associated with the touch can be determined. After that, you can decide on the application in Steps. Thereafter, at step 80, the current state (e.g., current mode) can be determined. Thereafter, in the above steps, the above-described ^ action set is set or selected according to the attributes of the above-mentioned _ 矣 、, m — - ’, 疋 (steps 74 to 80). For example, 'using the attribute of the decision, the system can refer to the stored gesture of the specific gesture set to the stored gesture map by default, or can be mapped by action: off::: :::: Use the control panel to change and articulate the gesture 4 values, or use the training (4) ning sequence to pass 128630.doc -24- 200847001. Figure 5 shows a 100 according to an embodiment of the present invention.柃 i i i 从 1 1 1 i i 1 1 1 1 1 1 1 1 1 刼 刼 刼 刼 刼 刼 刼 刼 于 于 于The touch is detected to the right, and the operation can continue to step 104 to identify the contact configuration. Step 104 can include substeps (4) 6 and (10). =

106中’決定是否可精確地識別接觸面積。例如 積是否係食指或拇指或手掌1無法精確識別,㈣ 繼續至步驟刚,在此步驟決定接觸面積數目。例如,: ^存在兩個接觸面積、三個接觸面積等等。步驟 疋 =可㈣至步驟11G ’在此步驟將辨識之接觸配置與 丁〜動作映射中儲存之接觸配置比較。若不匹配 可返回至步驟⑽。若存在匹配,則操作可繼續至= -,在此步驟在參考示意動作映射後,載入與初始;: 配f相關聯之示意動作集。之後,步驟U6中,可針對: 不思動作集相關聯之示意動作事件監視觸控。若實行示咅 動作事件,操作可繼續至步驟118,在此步驟實行與= 動作事件相關聯之命令。 Μ 圖6顯示根據本發明之—項具體實施例的說明性控制操 作120。控制操作12〇可始於步驟122,在此步驟摘測觸控 或接近觸控。之後,在步驟124,可針對觸控決定一弦。 之後,在步驟126 ’可決定與弦相關聯之輸入功能性。輸 入功能性可說明操作或操作群組。操作範例包括導覽操 ^、檔案操作、編輯操作、檢視操作、插人操作、格式化 操作、工具操作、網頁瀏覽操作等。輸入功能性可進一步 係基於%境狀況’包括觸控表面上之弦位置、應用、模式 128630.doc -25- 200847001 等’後I步驟128,可啟動或載入與輸入功能性相關106 determines whether the contact area can be accurately identified. For example, whether the product is the index finger or the thumb or the palm 1 cannot be accurately identified. (4) Continue to the step just before, and the number of contact areas is determined at this step. For example: ^ There are two contact areas, three contact areas, and so on. Step 疋 = (4) to step 11G ′ In this step, the identified contact configuration is compared with the contact configuration stored in the D to Action map. If it does not match, you can return to step (10). If there is a match, the operation can continue to = -, after this step is referenced to the action map, the load is associated with the initial;: the set of gestures associated with f. Then, in step U6, the touch can be monitored for: the gesture action event associated with the action set. If an action event is performed, the operation may continue to step 118 where a command associated with the = action event is executed. Figure 6 shows an illustrative control operation 120 in accordance with an embodiment of the present invention. Control operation 12 can begin at step 122 where the touch or proximity touch is taken. Thereafter, at step 124, a string can be determined for the touch. Thereafter, the input functionality associated with the chord may be determined at step 126'. Input functionality describes the operation or group of operations. Examples of operations include navigation operations, file operations, editing operations, viewing operations, insert operations, formatting operations, tool operations, web browsing operations, and the like. The input functionality can be further based on the condition of the % environment, including the position of the string on the touch surface, the application, the mode 128630.doc -25- 200847001, etc. After the I step 128, the bootable or loadable is related to the input functionality.

聯的示意動作隼。千咅&amp; A 不^動作集可為示意動作事件群組,其 可係指派給關於輸人功能性的不同動作。可藉由初始幻乍 為群,且群組之子集或相對於彼此地實行示意動作事 件。或者’可藉由與弦不相關聯之物件實行(例如在設定 基弦後添加之接觸面積)。示意動作事件可包括第一等級 芩數,例如運動、n敏 成擊、1Τ留、壓力及/或類似物。示音The gesture of the joint is awkward. The Millennium &amp; A No Action Set can be a group of action action events that can be assigned to different actions regarding the functionality of the input. The schematic action events can be performed by the initial illusion as a group, and a subset of the groups or relative to each other. Or ' can be performed by an object not associated with the chord (e.g., the contact area added after setting the basestring). The illustrated action event may include a first level of number of turns, such as motion, n-sensitive strike, 1 hold, pressure, and/or the like. Sound

動作事件亦可包括定㈣—等級減之第二等級參數,: :速=方向、形狀、時序/持續時間、長度及/或類似 &lt;,在步驟130’可在實行示意動作事件時實施盥 不思動作事件相關聯之動作。 〃 圖:示根據本發明之一項具體實施例的說明性示意動 控。步驟um: 在此步驟摘測觸 觸 呆.〇繼續至步驟144,在此步驟辨識接 觸配置。之後,步驟146 關聯的弦。步驟146德强 與辨識之接觸配置相 載入盘弦“可繼續至步驟148,在此步驟 何接觸配置在觸控期間實二包含可採用任 _,可針對藉由任何接觸 視觸控。若已葬由杯彳位 意動作事件監 可繼嘖至牛/ 7觸配置實行示意動作事件,操作 之控此步驟起始與示意動作事件相關聯 驟154,&gt; 止 “勡作事件,刼作可繼續至步 在此步驟決定是否仍偵 务 觸控,操作可返回步驟15〇。若 I ’右仍偵測到 哪右未偵測到觸控,操作可返 128630.doc -26 - 200847001 回至步驟142。即,抬起後跟觸控 意動作集。 “重-弦,、而重設示 圖7中之上述初始弦選擇模型有利地使使用 放下額外手指,而不影響示意動作集。此更合乎人體: I :因為以兩個手指之基弦開始的使用者可放下剩餘三個 Γ二以便所有五個手指支撐手。因此,可在較長時間週 』内“于示意動作。本質上’對基弦之改變被忽略。另 外,其使使用者可在有限空間觸控表面上藉由抬起除一手 ::之全部手指並朝目標前進及滑動其他手指而延 /拖曳、捲動等。 圖8顯示根據本發明 τ5 θ _ ^ # 不明之一項具體實施例的說明性示奄動 作操作1 60。摔作I μ w仏从止 ^ 耜作160可始於步驟162,在此步驟決定是否 伯測到2個相鄰手指。若是,操作可繼續至步驟⑹,在此 步驟根據動作實行指向操作。若否,操作166可繼續至步 驟162,在此步驟決定是否偵測到3個相鄰手指。若是 作可繼續至步驟168,在此步驟根據動作實行拖良操作。 若否,操作可繼續至步驟m,在此步驟決定是否㈣到 拇指及兩個相鄰手指。若是,操作可繼續至步驟172,在 此㈣實仃,要拖$操作。若否,操作可返回步驟162。 w序可在每_人所有手指從觸控表面抬起時(例如不再偵 測到觸控)予以重設。 、 圖9顯示根據本發明之一項具體實施例的說明性示意動 作#作180。操作18()可始於步驟182,在此步驟決定其 弦。之後,操作可分離或同時實行三個不同程序(平❹ 128630.doc -27- 200847001 驟184至ί88)。步驟184中,可偵洌 後,在步驟㈣弦之動作。之 中,可福、目,丨虹 只仃私向才呆作。步驟1 86 了偵刹新的第一手指。即The action event may also include a fourth level parameter of the (four)-level reduction, : speed = direction, shape, timing/duration, length, and/or the like, and may be implemented at step 130' when the action action is performed. Do not think about actions associated with action events. : Figure: Shows an illustrative schematic motion control in accordance with an embodiment of the present invention. Step um: At this step, the touch is touched. Continue to step 144 where the contact configuration is recognized. After that, step 146 associates the strings. Step 146: Deqiang and the identification contact configuration are loaded into the chord. "Continue to step 148. In this step, the contact configuration may include any _ during the touch period, and may be used for any touch-sensitive touch. The burial event is monitored by the cup and the action event can be continued to the cow/7-touch configuration to perform the action action event. The operation control is initiated in association with the action action event 154, &gt; You can continue to step at this step to decide whether to still touch the touch, the operation can return to step 15〇. If I ‘right still detects which right is not detected, the operation can return to 128630.doc -26 - 200847001 and return to step 142. That is, the heel is followed by the touch action set. "Heavy-chord," and resetting the initial string selection model described above in Figure 7 advantageously allows the use of additional fingers without affecting the set of gestures. This is more human: I: because the base of two fingers begins The user can drop the remaining three 以便 so that all five fingers support the hand. Therefore, it can be "scheduled" for a longer period of time. Essentially, changes to the basestring are ignored. In addition, it allows the user to extend/drag, scroll, etc. on a limited space touch surface by lifting all fingers except one hand and moving toward the target and sliding other fingers. Figure 8 shows an illustrative operation 1 60 of an embodiment of τ5 θ _ ^ # unknown according to the present invention. The fall of μ μ 仏 160 160 begins at step 162, where it is determined whether two adjacent fingers are detected. If so, the operation can proceed to step (6) where the pointing operation is performed in accordance with the action. If not, operation 166 may continue to step 162 where it is determined whether three adjacent fingers are detected. If yes, proceed to step 168 where the drag operation is performed according to the action. If not, the operation can continue to step m, where it is determined whether (4) to the thumb and two adjacent fingers. If so, the operation can proceed to step 172, where (4), the operation is to be dragged. If no, the operation returns to step 162. The w-order can be reset when all the fingers of each person are lifted from the touch surface (for example, no touch is detected). Figure 9 shows an illustrative schematic action #180 in accordance with an embodiment of the present invention. Operation 18() may begin at step 182 where the chord is determined. Afterwards, the operation can be separated or simultaneously implemented in three different programs (Ping Wei 128630.doc -27- 200847001, steps 184 to ί88). In step 184, the action of the string in step (four) can be detected. Among them, Kefu, Mu, and Yu Hong only stayed in private. Step 1 86 has a new first finger to detect the brake. which is

新第-手指(除基弦外)。之後,在步偵驟貝=弦不相關^ 測到新第-手指時起始第一命令二=可在每次須 藉由連續敲擊新第—手指實行重複中,使用者可 偵測到新第二手指(除基弦外)。即,^ ^驟MS中,可 聯的新第二手指。之後,…即可谓測與基弦不相關 新第二手m “驟194中’可在每次偵測到 、查碎 第二命令。某些情形中,使用者可夢由 連㈣擊新第…或簡單地 吏用::” 9内所述操作有時稱為當前弦選擇模型…丁重硬&quot;。圖 •Γ中之;範例中,基弦可為3個手指(食指-中指-無名指或 曰無名指)。因此’三個手指可用於 ΓΓ指已係基礎之部分,新第—手指可為拇指或食 :;二,拇指或食指可起始第—命令。另外,新第二手 日可為小才曰。因此,小指可起始第二命令。 —上述範例可在遊戲模式中很有用。基弦可用於指向,第 :新手指可用於觸發’第二新手指可用於武器改變。或 者’標準模式中’基弦可用於指向,第一新手指可用於主 要點選及拖良,第二報主-r ^ 弟一新手心可用於次要點選及拖曳。三手 指支援在遊戲及標準模式中均提供充分支援。 圖10』不根據本發明之一項具體實施例的說明性示意動 作才呆作200。操作200可始於步驟繼,在此步驟偵測觸 t v驟204後’操作可繼續至步驟,在此步驟辨識接 128630.doc -28- 200847001 觸配置。之後,步驟2〇6令,可決 關聯的弦。步驟20接 、/、辨識之接觸配置相 驟206後,知作可繼續 載入與弦相_之示意動作驟在此步驟 仃接觸配置在觸控 仗 驟210,m 丁之不思動作事件。之後,在步 叶對藉由任何接觸配置實 視觸控。若已藉由任何接觸配置動作事件監 可繼續至牛^ 明配置μ “意動作事件,操作 J、,黡、,貝至步驟212,在此 之控制/命令。若未^ _立起%不思動作事件相關聯 驟214,h T?F意動作事件,操作可繼續至步 鄉2 14 ’在此步驟法 ^ 期間暫停。若^ 接觸配置是否已在觸控 返回步驟204。若否,榀也π 絲作了 繼續至步驟216,在此步騾決 疋、測到觸控。若仍偵測到觸控,操作可返回至+ 細。若未偵測到觸控,操作可返回至步驟2〇2。 ^ 匕在圖10所述之暫停選擇模型中,切換僅可發生於最初手 扣暫侉或非常緩慢地移動之情況下。因此,使用者可從1New first-finger (except for the basestring). After that, in the step detection step = string is not relevant ^ when the new first-finger is detected, the first command is made = the user can be detected by repeating the new first-finger each time. New second finger (except for the base string). That is, in the MS, a new second finger can be connected. After that, ... can be said to be uncorrelated with the base string. The second second hand m "in step 194" can detect and check the second command every time. In some cases, the user can dream of (four) hit the new ...or simply use::" The operation described in 9 is sometimes referred to as the current string selection model... D. Hard &quot;. Figure • In the case; in the example, the base string can be 3 fingers (index finger - middle finger - ring finger or 曰 ring finger). Therefore, 'three fingers can be used for the part of the base that has been tied, the new one can be the thumb or the food: two, the thumb or forefinger can start the first command. In addition, the new second hand day can be a small talent. Therefore, the little finger can initiate the second command. - The above examples are useful in game mode. The base string can be used for pointing, the first: new finger can be used to trigger 'the second new finger can be used for weapon change. Or the 'standard mode' base string can be used for pointing, the first new finger can be used for the main point selection and the drag, and the second report main-r^ brother can be used for the next point selection and dragging. Three-finger support provides full support in both game and standard modes. Figure 10 is an illustration of an exemplary operation that is not in accordance with an embodiment of the present invention. Operation 200 may begin with a step after the step of detecting a step 204. The operation may continue to the step where the identification is made to 128630.doc -28-200847001. After that, step 2〇6 can be used to determine the associated string. After step 20 is connected to and/or identified by the contact configuration step 206, it is known that the gesture can continue to be loaded with the chord phase _ in this step 仃 contact is configured in the touch step 210, m does not think about the action event. After that, the touch is performed on the step by any contact. If any contact configuration action event has been monitored, you can continue to configure the μ "Important Action Event, Operation J,, 黡,, 贝 to Step 212, Control/Command here. If not ^ 立立%不The action action event is associated with step 214, h T? F is an action event, and the operation can continue until step 2 14 ' during this step method ^ pause. If ^ contact configuration has been touched back to step 204. If not, 榀Also, the π wire is continued to step 216, where the touch is detected and the touch is detected. If the touch is still detected, the operation returns to + fine. If the touch is not detected, the operation returns to the step. 2〇2. ^ In the pause selection model described in Figure 10, the switch can only occur when the initial buckle is temporarily paused or moved very slowly. Therefore, the user can go from 1

手指放鬆至五丰指扣A 于扣彳曰向,只要手指放下處於指向行程 間。 圖1 U員不根據本發明之一項具體實施例的說明性示意動 作操作220。操作220可始於步驟222,在此步驟決定是否 偵測到1個手指。若是,操作可繼續至步驟224,在此步驟 根據運動實行指向操作。若否,操作可繼續至步驟咖, 在此步驟決定是否偵測到2個相鄰手指。若是,操作可繼 續至步驟228,在此步驟根據動作實行拖曳操作。某些情 形中,可藉由點選起始拖曳鎖定/延伸,其他情形中,藉 128630.doc -29- 200847001 由放下拇&amp;起始(兩個相鄰手指+拇指)。若否 續至步驟230,在此步 知作可繼 • t ^ H , 疋6 1貝判到兩個不相鄰手 才曰:右疋,操作可繼續至步驟232,在此步驟實行次要拖 矣操作。某些情形中,可藉由 」精由點選起始拖曳鎖定/延伸, ::^ ,藉由放下拇指起始(兩個不相鄰手指+拇 指P若否,操作可繼續至步驟⑽,在此步驟 測到3或四個年沪。μ θ 义疋洛1貝Releasing your finger to the Wufeng fingertip A is in the direction of the buckle, as long as the finger is lowered and is pointing to the stroke. Figure 1 is a schematic illustration of an operational operation 220 that is not in accordance with an embodiment of the present invention. Operation 220 may begin at step 222 where it is determined whether one finger is detected. If so, the operation can continue to step 224 where the pointing operation is performed in accordance with the motion. If not, the operation can continue to the step coffee, in this step to determine whether two adjacent fingers are detected. If so, the operation can continue to step 228 where a drag operation is performed in accordance with the action. In some cases, the initial drag can be locked/extended by clicking on the point. In other cases, by 128630.doc -29- 200847001, the thumb &amp; start (two adjacent fingers + thumb). If it is not continued to step 230, in this step, it can be determined that t ^ H , 疋 6 1 Bay judged two non-adjacent hands: right, the operation can continue to step 232, in this step to implement a secondary Drag and drop operation. In some cases, you can start dragging the lock/extension by clicking on the selection, ::^, by lowering the thumb start (two non-adjacent fingers + thumb P if no, the operation can continue to step (10), In this step, 3 or 4 years of Shanghai was measured. μ θ Yiluo 1 shell

236)。若否:疋’可根據運動起始捲動(步驟 否,刼作可返回步驟222。此程序可在每次所 手指從觸控表面抬起時(例如不再制到 刻的情況下予以重設。 ’ 资月 圖12顯示根據本發明之一适且辨告Α 具體實施例的說明性示意動 呆 。刼作240可始於步驟242,在此步驟決定是否 偵測到5個手相。名:4 ,α &gt;&gt; _ ^ 和右疋,彳呆作可繼續至步驟244,在此步驟 起始指向’直至從觸控表面抬起所有手指。步驟⑽中, 决定5個手扣之一疋否在指向期間抬起及敲擊。若是,可 實行點選(步驟248)。步驟25〇中,決定5個手指中2個是否 在指向期間抬起及敲擊。若是,可實行拖,(步驟252”之 後,步驟253中,決定5個手指中1和2個是否在拖髮期間抬 起及破擊。若是,可實行放下(步驟⑸)。若否,拖良可繼 續。步驟250及254後,操作可返回步驟244。此程序可在 每,所有手指從觸控表面抬起時(例如不再偵測到觸控)和 暫A&quot;弦片刻的情況下予以重設。 圖丨2之替代具體實施例中 左之手指可起始左按紐點選 ’抬起及敲擊五手指基弦中向 ,抬起及敲擊基弦中向右之手 128630.doc •30- 200847001 指可起始右按鈕點選。 圖13顯示根據本發明之-項具體實施例的說明性示意動 作知作260。操作260可始於步驟262,在此步驟偵測觸 控。之後,在步驟264中,可辨識小指側面。小指側面產 生車乂長U度偏心)水平接觸面積,其不同於所有其他接觸 (除可能的扁平拇指外)。同樣,示意動作系統可能夠一致 地將水平偏心且薄接觸分類為小指(和可能的拇指)。之 後,在步驟266中,可谓測小指側面之擦過。之後,步称 268中’可實行與擦過之小指相關聯的控制和命令。 側面小指擦過可對簡單控制示意動作很有用,例如音量 =高/調低、睡眠、勞幕保護程式等。側面小指擦過具有 早一手指觸控之簡單性,同時可辨識地不同於正常丨指尖 指向/點選。同時亦非常合乎人體卫學,因為在採用小指 側面觸控日$ ’拇•曰豎直指向空中。此係所有腕部位置中最 中性舒適的。 圖14顯示根據本發明之一項具體實施例的說明性示意動 作刼作280。操作280可始於步驟282,在此步驟提供向上 滑動至公開及向下滑動至儀錶板之示意動作集映射。 儀錶板係包括可自訂界面工具集(微型應用程式)之控制 面板,其立即將資訊帶給使用者:天氣預報、證券報價、 黃頁、航班資訊、比賽分數等等。大多數情形中,當啟動 時’儀錶板開始檢視,當停用時,儀錶板移除檢視。使用 者能夠藉由按鈕點選從網際網路接收最新及及時資訊,然 後當鬆開按紐時其會立即消失。 128630.doc &gt;31 - 200847001 公開係視窗管理程式。視窗管理程式經組態心幫助濁 覽和減輕視窗雜波(由於開啟視t及/或應用程式太多而難 以找到文件和查看桌面之狀態)H可具有和係選自三236). If no: 疋 ' can be scrolled according to the motion start (step no, 可 can return to step 222. This procedure can be used each time the finger is lifted from the touch surface (for example, if it is no longer made to the moment) Fig. 12 shows an illustrative schematic diagram of a specific embodiment in accordance with one of the present invention. The operation 240 can begin at step 242, where it is determined whether five phases are detected. :4 , α &gt;&gt; _ ^ and right 疋, 彳 彳 可 可 可 可 可 可 可 可 可 可 可 可 可 可 可 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 244 If not, raise and tap during the pointing period. If yes, click to select (step 248). In step 25, determine whether 2 of the 5 fingers are raised and tapped during the pointing period. If yes, drag can be performed. After (step 252), in step 253, it is determined whether 1 or 2 of the 5 fingers are raised and broken during the twitching. If so, the plunging can be performed (step (5)). If not, the dragging can continue. After 254, the operation can return to step 244. This program can be used on every finger from the touch surface. Reset when lifting (for example, no touch is detected) and temporary A&quot; string. In the alternative embodiment of Figure 2, the left finger can start the left button and select 'Lift and Strike the center of the five-finger base string, lift and tap the right hand in the base string 128630.doc • 30- 200847001 refers to the start of the right button click. Figure 13 shows a specific embodiment in accordance with the present invention. The illustrative gesture is known as 260. Operation 260 can begin at step 262, where the touch is detected. Thereafter, in step 264, the side of the little finger can be identified. The side of the little finger produces a vehicular length U degree eccentricity) horizontal contact area, It differs from all other contacts (except for possible flat thumbs). Likewise, the gesture system can consistently classify horizontal eccentricity and thin contact as a little finger (and possibly a thumb). Thereafter, in step 266, the small finger can be measured. The side is wiped. After that, the step 268 can be used to control and command the associated little finger. The side finger can be used for simple control, such as volume = high / low, sleep, screen saver The side finger rubs the simplicity of the early finger touch, and is identifiably different from the normal fingertip pointing/clicking. It is also very ergonomic, because the little finger is used to touch the day. Vertically directed into the air. This is the most neutral of all wrist positions. Figure 14 shows an illustrative schematic action 280 in accordance with an embodiment of the present invention. Operation 280 may begin at step 282 where Provides a schematic action set mapping that slides up to the public and down to the dashboard. The dashboard includes a customizable interface tool set (micro-app) control panel that immediately brings information to the user: weather forecast, securities Quotes, yellow pages, flight information, match scores, and more. In most cases, the dashboard starts viewing when it starts, and when it is deactivated, the dashboard removes the view. The user can receive the latest and timely information from the Internet by clicking the button, and then disappears when the button is released. 128630.doc &gt;31 - 200847001 Open the window management program. The Windows Manager helps you to navigate and mitigate window clutter (difficult to find files and view desktop status due to too many open views and/or too many applications) H can have and select from three

種不同操作模式,其可藉由示意動作控制。第—模式係所 有視窗和並排、縮放及顯示全部。當在此模式中操作時, 並排及縮放所有開啟視f,以便可在顯示螢幕内同時查看 所有開啟視窗,,實行指定示意動作立即並排所有開啟 應用程式:縮小並整齊排列’以便使用者可檢視各視窗中 之内容。縮放數量和縮放率可與示意動作之某些特徵聯 結::如壓力和速度。第二模式係應用程式視窗和反白顯 不當W應用程式。此模式之工作方式類似於第一模式,除 了其僅在特定應用程式上卫作外。例如,實行指定示意動 作可立即並排特定應用程式之開啟視窗,同時使其他開啟 應用程式淡化至灰色陰影。第三模式係桌面或隱藏全部。 此模式中,將所有開啟視窗移動至螢幕邊緣,從而開放桌 面。即,實行指定示意動作可隱藏全部開啟視窗,從而使 使用者立即存取其桌面。 之後,步驟284中,操作可包括辨識基弦(例如兩個或兩 個以上手指)。之後,在步驟286中,決定是否偵測到基弦 之7上滑動。若是,操作可繼續至步驟288,在此步驟開 啟公開全部。之後,在步驟29〇中,決定是否偵測到基弦 之1下滑動。若是,操作可繼續至步驟292,在此步驟關 閉公開全部。之後,操作可返回步驟286。再參考步驟 286,右弦未偵測到向上滑動,則操作可繼續至步驟Μ*, 128630.doc -32- 200847001 在此步驟決定是否拍:目,丨5丨I &amp; 繼牛_ 彳到絲之向下_。若是,操作可 、.k、,、只至步驟296,在并牛驟p,热應力士上 在此步驟開啟儀錶板。之後,在步驟298 ’決定是否摘測到向上滑動。若是,操作可繼續至步驟 ’在此步驟關閉儀錶板。之後,操作可返回步驟施。 再多考步驟294,若未感測到向下滑動,操作可繼續至步 驟則,在此步驟決定弦是否已改變。例如,抬起」個手 心。右弦未改變’操作可繼續至步驟302,在此步驟採用Different modes of operation, which can be controlled by gestures. The first mode is all windows and side by side, zoom and display all. When operating in this mode, side-by-side and zoom all open view f so that all open windows can be viewed simultaneously in the display screen, and the specified gestures are performed immediately and all open applications are side-by-side: zoomed out and neatly arranged so that the user can view The content in each window. The number of zooms and the zoom ratio can be linked to certain features of the gesture: like pressure and speed. The second mode is the application window and the anti-white W application. This mode works in a similar way to the first mode, except that it is only available on a specific application. For example, performing a specified gesture can immediately side-by-side the opening of a particular application while allowing other open applications to fade to a shade of gray. The third mode is desktop or hidden. In this mode, all open windows are moved to the edge of the screen to open the desktop. That is, performing a specified gesture can hide all open windows, allowing the user to immediately access their desktop. Thereafter, in step 284, the operations may include identifying a base chord (e.g., two or more fingers). Thereafter, in step 286, it is determined whether or not the sliding of the base string 7 is detected. If so, the operation can continue to step 288 where all of the disclosures are opened. Then, in step 29, it is determined whether or not the base chord 1 is detected to slide down. If so, the operation can continue to step 292 where all of the disclosures are closed. Thereafter, the operation can return to step 286. Referring again to step 286, if the right chord does not detect an upward slid, the operation can continue to step Μ*, 128630.doc -32- 200847001. In this step, it is decided whether to shoot: 目, 丨5丨I &amp; Silk down_. If so, the operation can be, .k,,, and only to step 296, in the step of the heat stress, the dashboard is opened in this step. Thereafter, at step 298', it is determined whether or not the swipe is swiped up. If yes, the operation can proceed to step ’ at this step to close the dashboard. After that, the operation can return to the step. Step 294 is further tested. If the sliding down is not sensed, the operation can continue to the step where it is determined whether the string has changed. For example, raise your hand. The right chord has not changed' operation can continue to step 302 where it is employed

改變之弦實行指向。之後,在步驟3〇4中,決定弦是否已 改變回至基弦。 右疋刼作可返回步驟286。示意動作 作280有時稱為反向取消。 下雜 額外具體實施财,®14所^之料可係修改為包括向 右:向左擦過’其各與公開之不同功能性相關聯。 靜態命令示意動作係藉由簡單地在觸控表面上放置特定 =觸配置實行的示意動作。當在觸碰時識別特定接觸配置 時’起始-命令。初始放置後不存在示意動作事件。命令 ^在積測到觸控¥產生。靜態命令示意動作可類似於符號 語:’其中特定手配置意味著不同事情。靜態命令示意動 作背景中,各手配置(或不同接觸配置)起始-不同命令。 圖I5顯示根據本發明之一項具體實施例的說明性示意動 作#作310。不意動作操作31〇可始於步驟η〗,在此步驟 债測觸控表面上之觸控或接近觸控。步驟M2後,操作可 龜續至步驟3 14 ’在此步驟識別與觸控相關聯 配置。可透過-或兩隻手,或者替代或額外地透過= (例如觸控筆或圖章)完成接觸面積配置。當藉由手建立接 128630.doc • 33 - 200847001 觸面積時,可藉由手之任何部分來完成,包括手指、手掌 等,進一步係藉由手之任何方位,藉由手指相對於彼此之 任何位置,及藉由手指之任何方位。應明白,不同手方位 及不同手指位置/方位蔣逢&gt; τ k _ 直乃位將建立不同接觸面積配置。不同方 位之範例包括手張開朝下、手張開朝上、手侧面向下(拇 =向上),、手併攏-手掌向下、手併攏_手背、手併攏-指 節、手併撤-側面向下。;^ ρη主杜y 回门下不同手指位置之範例包括張開、 中性77、、且等彳預想任何接觸配置(符號語言)。 步驟Η4後,操作可繼續至步驟叩,在此步驟當識別特 定接觸配置時起始聯結至特定接觸配置之命令。命令可廣 泛變化。 作根據本發明之-項具體實施例的說明性示意動 作細作32〇。示意動作操作32〇可始於步驟如,在此步驟 針對動作監視觸控。例如, ^ ^七‘ 聆碉衩具體化為特定接觸配 。在運動’示意動作操作可繼續至步驟324,在此 步驟在輕敲與擦過間區分 換、Α 粍嘁可為快速較短運動。 U過可為較長較緩慢運動。 輕敲…… 胃t運動㈣破,則可在偵测到 Λ動作(步驟326)。若運動係擦過,則可在福 測到擦過時實行第二動作(步驟328)。 、 圖】7顯示根據本發明之一項且體 作操作330 -立“ H細例的說明性示意動 韦作330。不意動作操作33〇可始 谓測觸控表面(例如觸控板、觸控營 二此:驟 332後,操作可繼續至步驟334,在此牛j之觸控。步驟 聯之兩個接觸。&quot;驟識別與觸控相闕 步驟336中’操作可包括針對示音 128630.doc -34· 200847001 動作事件監視觸控。步驟 步驟338中,當第一接觸係固定而第 -接觸繞第-接觸旋轉時可起始旋轉命令。 2使用者在任—時間保持多個視f開啟。此允許他們 根據需要在視窗間來回。♦ 口 田攸一視窗複製並且另一視窗内 貼上時,此點可很右剎 ° 〃 1L程序中,使用者將選擇視窗 之一,從而啟動視窗,麸έ …、後在視自内實行動作。使用者接 著會選擇另一視窗,户 ^ _攸而啟動新視窗並停用舊視窗,然後 在新視窗内實行動你 lL ^Change the string to implement the direction. After that, in step 3〇4, it is determined whether the string has changed back to the basestring. Right hand can return to step 286. The gesture 280 is sometimes referred to as a reverse cancellation. In addition to the additional specific implementation, the materials of the 14 can be modified to include right: rub to the left' each of which is associated with a different functionality of the disclosure. The static command gestures the gesture by performing a specific = touch configuration on the touch surface. When starting a specific contact configuration when touching, the 'start-command'. There is no indicative action event after the initial placement. The command ^ is generated in the product test to touch ¥. A static command gesture can be similar to a symbol: 'where a particular hand configuration means a different thing. The static command indicates the start-different command for each hand configuration (or different contact configuration) in the context of the action. Figure I5 shows an illustrative schematic action #310 in accordance with an embodiment of the present invention. Unintentional operation 31 can start at step η, where the touch or proximity touch on the touch surface is measured. After step M2, the operation may continue to step 3 14 ' at this step to identify the configuration associated with the touch. The contact area configuration can be done through - or both hands, or alternatively or additionally via = (such as a stylus or stamp). When the contact area is established by hand, it can be done by any part of the hand, including the fingers, the palms, etc., further by any position of the hand, by means of any of the fingers relative to each other. Position, and any orientation with your fingers. It should be understood that different hand orientations and different finger positions/orientations of Jiang Feng&gt; τ k _ straight will establish different contact area configurations. Examples of different orientations include hand open, hand open, hand side down (bump = up), hands close together - palm down, hands close together _ hand back, hands close together - knuckles, hands and sides - side down. ;^ ρηMain Duy Examples of different finger positions under the door include open, neutral 77, and etc. Expect any contact configuration (symbolic language). After step Η4, the operation can continue to step 叩, where the command to link to a particular contact configuration is initiated when a particular contact configuration is identified. Commands can vary widely. An illustrative schematic action of a specific embodiment in accordance with the present invention is detailed. The gesture action 32 can begin with a step such as, in this step, monitoring the touch for the action. For example, ^^7 ‘the hearing is embodied as a specific contact. In the motion&apos; gesture action may continue to step 324 where the tapping and rubbing may be a quick and short motion. U can be longer and slower. Tapping... Stomach t motion (4) breaks, then a squat action can be detected (step 326). If the exercise is swiped, the second action can be performed when the wipe is detected (step 328). FIG. 7 shows an exemplary operation of the apparatus according to one embodiment of the present invention. The display of the touch surface (for example, the touch panel and the touch) can be started. Control Camp 2: After step 332, the operation may continue to step 334, where the touch is touched. The steps are linked to the two contacts. & </ br><br><br><br><br><br> 128630.doc -34· 200847001 Action event monitoring touch. In step 338, the rotation command can be initiated when the first contact system is fixed and the first contact is rotated around the first contact. 2 The user maintains multiple views during the time-time f. This allows them to go back and forth between windows as needed. ♦ When a window is copied and another window is attached, this point can be very right. 〃 In the 1L program, the user will select one of the windows to start. Window, bran ..., after the action in the view. The user will then choose another window, the user ^ _ 攸 and start a new window and disable the old window, then act in the new window you lL ^

、 。匕較為笨拙,且由許多步驟組成。 馬克服此點,本發明捭版 &amp;从供®18之方法,其為使用者提供在 視囪間來回之能六,二—&amp; …、為啟動及停用。當在視窗上方定 位指向機制時視窗作用。 ^ 才曰向機制在觸控板情形中可為指 游標,或者在觸控螢幕情形中可為手指。 圖_示根據本發明之-項具體實施例的說明性示意動 如呆作340。不意動作操作340包括提供多個視窗342。例 如’視窗可為桌面上夕„私如』 之開啟程式。之後,步驟344中,操 作可包括偵測第一#窑μ +杜人他 見自上之私向機制。例如,在觸控板情 二/日向機制可為游標。在觸控螢幕情形中,指向機制 二手指或手指群㉞。之後,步驟地中,可監視觸控以 中,:、曰向機制相關聯之弦是否已存在。在觸控板情形 _ : 1橾放置於視窗上後完成此點。在觸控螢幕情 + 可在將手指放置於視窗上的同時完成此點。之後, 乂驟348中’可起始與弦及第—視窗相關聯之示意動作集 ^如示意動作集取決於弦及視窗)。之後,步驟35〇中,當 觸控表面上實行與示意動作集相關聯之示意動作事件 I28630.doc -35- 200847001 時’可在第一視窗内實行動作 可包括谓測第二視窗上之 ,:驟352中’操作 344之時間完成,或者 ?在不同於步驟 可為❹第-視時發生。指向機制 於“自之相同指向機制(例如將指向機制從_ 4視自移動至另一個),或, . It is awkward and consists of many steps. To overcome this, the present invention is directed to the method of &lt;18&gt;, which provides the user with the ability to switch back and forth between the viewing and closing chambers, six, two, & ..., for activation and deactivation. The window acts when the pointing mechanism is positioned above the window. ^ The mechanism can be referred to as a cursor in the case of a touchpad or as a finger in the case of a touch screen. Figure - shows an illustrative representation of a particular embodiment in accordance with the present invention. Unintentional action operation 340 includes providing a plurality of windows 342. For example, 'the window can be the opening program of the desktop on the eve of the private. Then, in step 344, the operation may include detecting the first # 窑 μ + Du Ren he sees the private mechanism of the top. For example, in the touchpad The second/day mechanism can be a cursor. In the case of a touch screen, the pointing mechanism is two fingers or a finger group 34. After that, in the step, the touch can be monitored, and the chord associated with the mechanism is already present. In the case of the touchpad _ : 1 橾 after placing it on the window to complete this point. In the touch screen + can be done while placing the finger on the window. After that, in step 348 'start and The schematic action set associated with the string and the first window is such that the set of action actions depends on the string and the window. Then, in step 35, the schematic action event associated with the set of action actions is performed on the touch surface I28630.doc - 35- 200847001 When the action can be performed in the first window, the second window can be included in the pre-measurement: in step 352, the time of operation 344 is completed, or when it is different from the step, it can occur as the first-view. Mechanism (E.g., a mechanism from point of view _ 4 from moving to another), or

及者扣向機制可為第二指向機制 (例如弟二游標或第一车拉、 或市J … 手'曰)。之後,步驟…中,可監視觸 弟一扣向機制相關聯之弦是否已存在。之 後,步驟356中,可起始與 汉弟一視自相關聯之示意動 作木。之後’步驟358中,各在觴批本品L虫 从隹上 田在觸控表面上實行與示意動 作集相關聯之示意動作事 叮 』隹罘一視自内實行動 作0 、广《之一乾例中’其使用觸控螢幕,若使用者正在實行 複製及貼上Μ吏用者可放置第—視窗上之_ +的手#,並 選擇欲複製之物件。之後,使用者可在第一視窗二于複 製不意動作。使用者亦可放置第二視窗上之第二手的手The deduction mechanism can be a second pointing mechanism (for example, the second vernier or the first car, or the city J...hand 曰). After that, in step..., it is possible to monitor whether the chord associated with the mechanism is already present. Thereafter, in step 356, a schematic action tree associated with the Han brother may be initiated. Then, in step 358, each of the L-worms in the batch is executed on the touch surface by Sakagada, and the gestures associated with the gesture set are implemented. In the case of using the touch screen, if the user is performing copying and pasting, the user can place the _+ hand# on the first window and select the object to be copied. After that, the user can perform the unintentional action in the first window. The user can also place the second hand on the second window.

指,並選擇用於貼上之位置。之後’使用者可在第二視窗 上實行貼上示意動作。 圖18之另一範例中,其使用觸控螢幕,若使用者正在實 行複製及貼上,使用者可放置第一視窗上之一手的手指, 並選擇欲複製之物件。之後,使用者可在第一視窗上實行 複製示意動作。使用者接著可將其手移動至第二視窗。使 用者可將手指放置於第二視窗上,並選擇用於貼上之位 置。之後,使用者可在第二視窗上實行貼上示意動作。 圖19顯示根據本發明之一項具體實施例的說明性示意動 128630.doc -36- 200847001 乍操作360不思動作操作可始於步驟364,在此步驟辨識 與第手相關聯之接觸配置。步驟364後,操作可繼續至 步驟=6 ’在此步驟载人基於與第—手相關聯之接觸配置 的丁心動作集。步驟366後,操作可繼續至步驟⑽, t驟辨識與第二手相關聯之第二接觸配置。步驟368後, 士呆作可繼、_至步Up 2 。 如 、 驟370,在此步驟根據第二接觸配置改變 及修改第一接觸配置之示意動作集。應明白,第二手可用 一】而要在不同不意動作集間迅速調變。此亦可提供更 見:圍之可能示意動作集。例如,若第一手可根據手指數 目=供⑽示意動作集’並且第二手可根據手指數目提化 個t改#,則可建立1〇〇個示意動作集。 ,、 引用圖Μ之範例,其使用觸控板,使用者可 :置:觸控板上,從而起始與兩個手指相關聯之示意I; 吏Γ可使用兩個手指實行示意動作事件。使用者可 因此使用者可放下來自相反Γ動供的動作。 由第一…f相反手之兩個手指。此可載入欲藉 由弟手只仃之全新示意動作集。 圖:顯示根據本發明之一項具趙實施例的說 =:操作可始於步驟3&quot;,在此步驟提供滑鼠及觸 :表例如’可將滑鼠連接至膝上型電腦,其 後,操作可繼續至步驟384,在此步驟針對滑 :事牛…見滑鼠(例如移動游標、按-點選 %,步驟386中,可針對示意動作事件監 同 驟388中,可根據滑鼠事件 二、面。步 只们月以呆作。同時,步驟390 I28630.doc -37- 200847001 中,可根據示意動作事件實 許使用者用一手控制滑鼠, 因此使用者可更具生產力。 行示意動作操作。此方法可允 同時用相反手實行示意動作。 圖20之一範例中,試 I 0〜、 者耘式申,使用者可使用滑鼠選 70,然後用相反手在觸控表 #田j a — η工衣面上實仃複製不意動作。 使用者接者可使用捲動哎-立 &lt;干移不思動作捲動或透過試算表 平移。一旦發現期望位, 使用者可使用滑鼠選擇用於貼 上IJ里單元,缺接與^Refer to and select the location to use for the label. After that, the user can perform a gesture on the second window. In another example of Fig. 18, a touch screen is used. If the user is copying and pasting, the user can place a finger of one hand on the first window and select the object to be copied. After that, the user can perform a copy gesture on the first window. The user can then move his hand to the second window. The user can place his finger on the second window and select the location to be used for the sticker. After that, the user can perform a gesture on the second window. Figure 19 shows an illustrative schematic diagram in accordance with an embodiment of the present invention. 128630.doc - 36 - 200847001 Operation 360 Operation may begin at step 364 where the contact configuration associated with the first hand is identified. After step 364, the operation can continue to step = 6 ' at this step to load the set of actions based on the contact configuration associated with the first hand. After step 366, the operation may continue to step (10), identifying a second contact configuration associated with the second hand. After step 368, the stalker can continue, _ to step Up 2. For example, step 370, in this step, the set of schematic actions of the first contact configuration is changed and modified according to the second contact configuration. It should be understood that the second hand can be used to change quickly between different sets of actions. This also provides a more visible: a set of possible gestures. For example, if the first hand can indicate the action set according to the hand index = (10) and the second hand can refine the t change according to the number of fingers, one set of gestures can be established. An example of a reference figure, which uses a touchpad, can be placed on the touchpad to initiate a gesture I associated with two fingers; 示意 two gestures can be used to perform a gesture event. The user can therefore drop the action from the opposite sway. Two fingers by the first...f opposite hand. This can be loaded with a new set of gestures that you want to borrow from your brother. Figure: shows an embodiment according to the present invention. =: Operation can begin in step 3 &quot;, in this step, provide a mouse and touch: a table such as 'can connect the mouse to the laptop, then The operation may continue to step 384, in which the step is directed to the slide: see the mouse (for example, moving the cursor, pressing - click on %, in step 386, for the gesture event can be monitored in step 388, according to the mouse Event 2, face. Steps are only a month to stay. At the same time, in step 390 I28630.doc -37- 200847001, the user can control the mouse with one hand according to the schematic action event, so the user can be more productive. Indicates the action operation. This method allows the gesture to be performed with the opposite hand at the same time. In one example of Figure 20, the test I 0~, the 耘 申 application, the user can use the mouse to select 70, and then use the opposite hand on the touch meter #田ja — η 工 衣 仃 仃 仃 仃 仃 η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η η The mouse is selected for attaching the IJ unit, missing ^

…傻貝仃貼上示意動作。 圖21顯示根據本發明之一 項/、體實施例的說明性控制摔 =控::〗操作可始”驟402,在此步驟提供滑鼠二 對二事:驟4〇2後,#作可繼續至步驟404,在此步驟針 门 件監視滑鼠(例如移動游標、独點選、捲動)。 驟406中’可針對示意動作事件監視觸控表面。 之後ν驟4〇8中,滑鼠模式可奸舍 —立 、飞了根據在觸控表面實行的觸 控或不思動作事件改變 主η 、 ^例如,觸控手可用於修改以另一 手操作之滑氣的功能性 κ 例如,觸控表面上手指之存在可 將滑鼠從追_式改變轉動模式。 作4回10。d根據本發明之一項具體實施例的說明性控制操 第一觸抻呆可始於步驟412 ’在此步驟偵測觸控表面上之 二Γ接步驟412後,操作可繼續至步驟414,在此步驟 辨識弟一接觸配詈。 例如,此可用第一手完成。同時,步 驟416及418中,可福:則總 w、觸控表面上之第二觸控,並且可辨 識弟一接觸配置。例 】如,此可用第二手完成。步驟420 入用於第一接觸配置之示意動作集。同時,步驟 128630.doc -38- 200847001 422中,可哉λ m 於第二接觸配置之示 中,可針對筮一九表 “切作集。步驟424 T野弟一不意動作事件監視第一觸 426中,可針對—立 啁控。同時,步驟 t弟一不忍動作事件監視二 中,可在實行第一 一觸控。步驟428 仃弟不思動作事件時實施盥M 一九去 件相關聯之動你 n &lt;、弟一不意動作事 p之動作。同時,步騾43〇中,可在 動作事件時實施盥篦一一立 、仃弟一不意 H23?g 4 —以、動作事件相_之動作。 圖23顯示根據本發明之一項 作440。摔作44f)7&amp;伙止 貝匕1夕』的呪明性控制操... silly bellows posted a gesture. Figure 21 shows an exemplary control of the fall control in accordance with one or another embodiment of the present invention. The operation may begin at step 402. In this step, the mouse is provided with two pairs of things: after the step 4〇2, #作Proceed to step 404, in which the needle member monitors the mouse (eg, moving the cursor, clicking, scrolling). In step 406, 'the touch surface can be monitored for the gesture event. Then, in step 4〇8, The mouse mode can be used to make a difference. The main η is changed according to the touch or non-thinking event performed on the touch surface. For example, the touch hand can be used to modify the functionality of the slippery operation with the other hand. The presence of a finger on the touch surface can change the rotation mode of the mouse from the chasing mode. 4 times 10. d. The illustrative control operation according to an embodiment of the present invention may begin at step 412. After detecting the second connection step 412 on the touch surface, the operation may continue to step 414 where the identification of the contact is identified. For example, this can be done with the first hand. Meanwhile, steps 416 and 418 are performed. In the middle, the blessing: the total w, the second touch on the touch surface, and For example, this can be done with the second hand. Step 420 is used to set the schematic action set for the first contact configuration. Meanwhile, in step 128630.doc -38- 200847001 422, 哉λ m is in the first step. In the description of the two-contact configuration, the set can be cut for the nineteen-nine table. Step 424 T the wilderness does not care about the event event monitoring the first touch 426, which can be directed to the vertical control. At the same time, the step t brother can't bear the action event monitoring 2, and the first touch can be implemented. Step 428 When the younger brother does not think about the action event, the implementation of the 盥M 九 去 件 件 相关 相关 相关 相关 相关 n n n 、 、 、 、 、 、 、 、 、 、 、 、 、 At the same time, in step 43, you can implement the action of the event, the action of the action, and the action of the event. Figure 23 shows a work 440 in accordance with one embodiment of the present invention. Falling 44f) 7&amp;

餘作440可始於步驟442, 之後,步驟444中,可在顧- _ ° 指卜^ UI物件之不同•則貞測多個 觸㈣幕情形中,第-指標可為第-手之手指, 為弟一手之手指。在觸控板情形中,第-指標 可為糟由第一手护制夕被挪 ^ ’、 弟二指標可為藉由第二手控 ㈣- 4,可在步驟445中將指標鎖定 於”、、員不之物件。此可藉由暫 6 凡成。例如,使用者可將指 ‘放置於⑽物件上之點’然後停留直至鎖定指標。之 灸V驟446中’可監視指標相對於其鎖定位置之位置。 之後,步驟448中,當相對於其鎖定位置移動一或多個指 標之位置時’可修改顯示之物件。例如’可修改⑽物件 之屬性,例如其包括方位、尺寸及形狀。 引用圖23之範例,為旋轉物件,使用者可將其指標放置 =GUI物件±,當鎖定時,旋轉一或兩個指標。⑽物件 可遵循一或兩個指標之旋轉,從而改變其方位。為改變其 尺寸,使用者可將指標放置於Gm物件上,當鎖定時,將 兩個指標滑動至一起或彼此遠離。當滑動至一起時,可減 128630.doc -39- 200847001 小物件大小。當滑動分開時,可增加物件大小。 形狀,使用者可將指標放置於GU 夂- 一個指標向上、向下以“真 田鎖疋時’將 下、向右、向左岣動。當從物 開時’物件在滑動方向上可變大。當向物件滑動時::: 在滑動方向上可變小。 牛 圖Μ顯示根據本發明之―項具財施例的說明 Π:作可始於步驟452,在此步驟_觸控表二 控。例如,此可用第一手完成。步驟452後,摔作 可=Γ454,在此步驟辨識與第-觸控相關聯3 = ΓΓ,步驟456及458中,測觸控表面: 弟:觸控,並且可辨識第二接觸配置。例如,此可用第 一手完成。步驟46G中,可針對第—示意動作事件監視第 一觸控。同時,步驟46中, - ^ 視第二觸控―,可= = 與第-示意動作事件相關聯之指向操作。同時,步: 可在實仃第二示意動作事件時實施與第二示竟 作事件相關聯之指向操作。 圖25顯示根據本發明之一項具體實施例的說明性控制操 作獨。控制操作可始於步驟472,在此步驟顯示可選擇項 目之:組。例如’可選擇項目之群組可為調色板或工呈 ^擇二驟Μ後,操作可繼續至步驟474,在此步驟痛測可 Μ撒目之群且上的第一指向機制。在觸控螢幕情形中, 二為為手指或手指群组。在觸控板情形中,指向機 為精由手指或手指群組控制之游標或工具。步驟476 128630.doc 200847001 中,動(選擇)第-指向機制下方之項目。啟動項目可 係基於扣向機制之位置。可提供 、 、 認其選擇。例如,當第心見胃讀使用者確 一 *弟寺日向機制位於項目上時,可 顯示啟動項目。步驟4 7 8 Φ,-Γ η + 中可同時偵測視窗上之第二於 =機制(例如兩個指向機制同時運作)。例如 = 内之帆布。觸控瑩幕情形中,第二二The remainder 440 may begin at step 442. Thereafter, in step 444, the difference may be in the case of a plurality of touch (four) scenes, and the first indicator may be the finger of the first hand. For the younger brother's finger. In the case of the touchpad, the first indicator can be moved by the first hand, and the second indicator can be controlled by the second hand (four)-4, which can be locked in step 445. For example, the user can place the finger 'at the point on the (10) object and then stay until the index is locked. The moxibustion V 446 can't monitor the indicator relative to The position of the locked position. Thereafter, in step 448, the item being displayed may be modified when the position of one or more indicators is moved relative to its locked position. For example, the attributes of the item may be modified (10), for example, including orientation, size, and Referring to the example of Figure 23, for rotating objects, the user can place their index = GUI object ±, when locked, rotate one or two indicators. (10) The object can follow the rotation of one or two indicators, thereby changing its Orientation. In order to change the size, the user can place the indicator on the Gm object. When locking, the two indicators are slid together or away from each other. When sliding together, the object can be reduced by 128630.doc -39- 200847001 Size. When slipping When separated, increased size of the object shape, the user may be placed in GU Wen index - an indicator up, down to "lock piece goods when Sanada 'lower, right, left Gou move. When the object is opened, the object can become large in the sliding direction. When sliding to the object::: It can be small in the sliding direction. The description of the item according to the present invention is shown in Figure 452. In this step, the touch table is controlled. For example, this can be done with the first hand. After step 452, the fall can be Γ 454, in this step, the identification is associated with the first touch 3 = ΓΓ, in steps 456 and 458, the touch surface is measured: the touch: and the second contact configuration can be recognized. For example, this can be done with the first hand. In step 46G, the first touch can be monitored for the first-illustrated action event. Meanwhile, in step 46, - ^ depending on the second touch, can = = a pointing operation associated with the first-illustrated action event. At the same time, step: The pointing operation associated with the second active event can be implemented when the second schematic action event is implemented. Figure 25 shows an illustrative control operation in accordance with an embodiment of the present invention. Control operations may begin at step 472, where a selectable item: group is displayed. For example, after the group of selectable items can be a palette or a second step, the operation can continue to step 474, where the pain can measure the first pointing mechanism on the group. In the case of a touch screen, the second is a finger or group of fingers. In the case of a trackpad, the pointing machine is a cursor or tool that is controlled by a finger or group of fingers. In step 476 128630.doc 200847001, move (select) the item below the first-pointing mechanism. The startup project can be based on the location of the deduction mechanism. Available, and available. For example, when the first-time stomach reader is sure that the diver's day-to-day mechanism is on the project, the startup project can be displayed. Step 4 7 8 Φ, -Γ η + can simultaneously detect the second on = mechanism on the window (for example, the two pointing mechanisms operate simultaneously). For example = canvas inside. In the case of touch screen, the second two

圖26顯不根據本發明之一項具體實施例的說明性控制揭 作遍。控制操作可始於步驟如,在此步驟顯示媒體項目 =群、i例如,媒體可為歌曲、視訊剪輯、電影、照片、 遊戲等等f列如,媒體項目可由媒體管理程式控制。由 CUPertln〇之Apple Inc.製造的iPhoto及iTUnes係媒體管理程 式之範例。步驟5〇4中,可㈣項目上之指標機制,從而 反白顯7F項目。例如’可藉由滑鼠或手弦(慣用手)控制指 標機制。步驟506中,可㈣觸控表面上之觸控。可藉由 觸控板情形中,第二指向機制可為藉由手指或手指群^ 制之游標或工具。步驟480後,可將選定項目應用於第二 指標機制。例如,在工具列情形中,第二指標機制之功能 性可基於選定項目。另夕卜’在觸控板情形中,工具U!元件 可隨功能性改變,以便❹者知道何^具作用(例如^ 筆、畫筆等)。在調色板情形中,第二指標機制之輸出可 基於敎項目。例如,繪畫情形中,當使用者選擇調色板 上之1^色日卞’第二指標之輸出改變至該顏色。因此,使 用者可在繪畫操作期間即時調整應用於帆布之顏色。 128630.doc -41 - 200847001 手貝仃觸控,其與指標機制相關聯之手相 驟508,可辨塒盥魅〜, I便,在步 中,可將耳^ 之接觸配置。之後,步驟51〇 目識之接觸配置的Μ料插人醒目顯示項Figure 26 shows an illustrative control disclosure in accordance with an embodiment of the present invention. Control operations may begin with steps such as displaying media items = groups, i, for example, media may be songs, video clips, movies, photos, games, etc., such as, media items may be controlled by the media management program. An example of iPhoto and iTUnes media management programs manufactured by Apple Inc. of CUPertln〇. In step 5〇4, the indicator mechanism on the project can be (iv), so that the 7F project is highlighted. For example, the indexing mechanism can be controlled by a mouse or a hand string (dominant hand). In step 506, the touch on the touch surface can be (4). In the case of a touchpad, the second pointing mechanism can be a cursor or a tool that is controlled by a finger or a group of fingers. After step 480, the selected item can be applied to the second indicator mechanism. For example, in the case of a toolbar, the functionality of the second indicator mechanism can be based on the selected item. In addition, in the case of a touchpad, the tool U! component can be changed with functionality so that the latter knows what effect (e.g., pen, brush, etc.). In the case of a palette, the output of the second indicator mechanism can be based on the item. For example, in the case of painting, when the user selects the color of the 1st color on the palette, the output of the second indicator changes to the color. Therefore, the user can instantly adjust the color applied to the canvas during the drawing operation. 128630.doc -41 - 200847001 Hand-held touch, its hand-related mechanism with the indicator mechanism, step 508, can identify the enchantment~, I will, in the step, the contact of the ear can be configured. After that, in step 51, the contact configuration of the contact is inserted into the eye-catching display item.

有第一…了具有不同兀貢料。例如,單-手指可具 :貝科,兩個手指可具有第二元資#,三個手指可 八弟二凡賁料,四個手指可具有第四元資料,五個手指 可具有弟五元資料,依此類推。用於各接觸配置之元資料 可係關於相同功能性,或其可具有完全不同之功能性。貝科 引用圖26之範例,上述程序可用於額定或分級媒體項 目°例如’單一手指可將1星分級聯結至歌曲’兩個手指 可將2星分級聯結至歌曲,三個手指可旧星分級聯結至歌 曲’四個手指可將4星分級聯結至歌曲,五個手指可將5星 分級聯結至歌曲。 引用圖26之另一範例,上述程序可用於識別媒體項目。 例如,單一手指可將第一人(例如妻子Darcy)聯結至照片, 第一手私可將第二人(例如女兒Halley)聯結至照片,第三 手才曰可將第二人(例如表兄Camden)聯結至照片,第四手指 可將第四人(例如叔父Chance)聯結至照片,第五手指可將 第五人(例如奶奶shery)聯結至照片。敲擊可用於放大清 單例如,第一觸碰之手指數目可被添加至第二觸碰之手 才曰數目,以產生最終接觸配置。三個手指後跟兩個手指可 仍意味著第五人。而五個說之及一個手指(總共6個手指)可 將第六人(例如曾祖母TenT)聯結至照片。另外,五個手指 及兩個手指可將第七者(例如狗Prima)聯結至照片,五個手 128630.doc -42- 200847001 :::三個手指可將第八人(朋友叫聯結 ^及四個手指可將第九人聯結至照片,而二 手指可趑结, 儿口于扣及五個 關聯,&quot;:人:二至照:。或者,,可與-人不相 景、家庭^ 某一其他特徵(例如假期、加州、風 一般而言, 式。傳統上, 聯之弦完成。 可需要將媒體㈣程式放置於元f料***模 此可透過螢幕上控#或藉由與指向機制相關There is a first... there are different tributes. For example, a single-finger can have: Beike, two fingers can have a second element, three fingers can be eight brothers, four fingers can have fourth yuan data, five fingers can have a fifth child Metadata, and so on. The metadata used for each contact configuration may be about the same functionality, or it may have completely different functionality. Referring to the example of Figure 26, the above program can be used for rated or graded media items. For example, 'single finger can link 1 star to songs' two fingers can link 2 stars to songs, and three fingers can be graded by old stars. Linked to the song 'Four fingers can link 4 stars to songs, and 5 fingers can link 5 stars to songs. Referring to another example of Figure 26, the above procedure can be used to identify media items. For example, a single finger can link a first person (such as a wife Darcy) to a photo, a first hand can link a second person (such as daughter Halley) to a photo, and a third hand can connect a second person (such as a cousin) Camden) is linked to the photo, the fourth finger can connect a fourth person (such as uncle Chance) to the photo, and the fifth finger can link the fifth person (such as grandma shery) to the photo. The tap can be used to zoom in on the list, for example, the number of fingers of the first touch can be added to the number of second touch hands to produce the final contact configuration. Three fingers followed by two fingers can still mean the fifth person. Five of them, one finger (a total of six fingers), can link a sixth person (such as the great-grandmother TenT) to the photo. In addition, five fingers and two fingers can link the seventh party (such as dog Prima) to the photo, five hands 128630.doc -42- 200847001 ::: three fingers can be the eighth person (friends call the connection ^ and Four fingers can connect the ninth person to the photo, while the two fingers can be knotted, the child's mouth is buckled and five related, &quot;: person: two to the photo: or,, can be different from the person, family ^ Some other feature (such as holidays, California, wind, style. Traditionally, the chord is completed. You may need to place the media (4) program in the meta-injection mode. This can be done via the screen control# or by Pointing mechanism related

7員不根據本發明之一項具體實施例的說明性控制操 。控制操作可始於步驟522,在此步驟顯示控制物 :°例如’控制物件可為屏幕上独、撥盤、滑塊、開關 等。步驟524中,可偵測控制物件上之觸控。之後,在步 驟526,可辨識與觸控相關聯之接觸配置。步驟527中,可 決定=觸配置是否匹配與控制該控制物件相關聯之接觸配 置°若不匹配’操作可繼續至步驟528,在此步驟忽略觸 控。若匹配,操作可繼續至步驟53〇,在此步驟針對與控 制该控制物件相關聯之示意動作事件監視觸控。若實行示 意動作事件,可將與其相關聯之動作應用於控制物件,以 便控制物件可產生其命令(步驟43 2)。 圖27之一項具體實施例中,起始控制該控制物件之接觸 配置可基於多個接觸,而忽略之接觸配置可基於單一接 觸。 圖28顯不根據本發明之一項具體實施例的說明性控制操 作540。可使用控制面板,例如圖μ所示者,實行控制操 128630.doc -43· 200847001 作540。操作可始於步驟%,在此步驟提供輸人功 單。輸入功能可具有命令及連結至命令之示意動作/月 命令可係關於輪入功能。例如,若輸入功能係編輯,心 令可為取消、復原、貼上、後退標籤、複製、標籤、; 下、取消復原及選擇全部。示意動作事件可為藉由弦實行 之動作。示意動作事件可包括旋轉(順時針及逆時、 為群組轉換(向上、向下、向右、向左)、相對於彼此轉換 (張開、併攏)、敲擊等。可以直觀方式將示意動作事件連 結至命令。例如’可藉由將手指併攏在一起,類似於剪下 動作,起始剪下命彳。另外,互補命令(剪下/貼上)通常係 連結至互補示意動作事件(張開/併攏)。步驟544中,可將 輸^ 力能指派給特定弦。弦可為手指配置。可將弦顯示於 ^月早内’以便使用者反白顯示特定弦,’然後選擇特定輸入 功能。可將弦指派給多個輸入功能。 步驟544中’使用者可藉由從輸人功能之選單選擇輸入 魏,然後從弦之選單清單選擇弦,將輸入功能指派給特 ^弦^例如反白顯示輸人功能’然後選擇弦)。一旦選擇 弦可將輸入功月匕聯結至該弦以供未來操作(步驟⑽)。某 些情形中’弦選單可包括指派給其之輸入功能的指示。此 特定具體κ &amp;例中,已將命令指派給示意動作事件。 圖29A至G係根據本發明之一項$體實施例的控制面板 550之圖式。圖29A顯示基本控制面板,而圖29B至29G顯 示項控制面板具體實施例之數個範例。 如圖29A所示,控制面板55〇可包括下拉視窗形式之示意 128630.doc -44- 200847001 動作映射控制552。當使用者選擇下拉視窗時,可為使用 者提供m㈣意動作映射選項。可為制者提供預設 没定值(通常由製造商設定)’或者使用者可選擇自訂設定 值’以便其可形成自己的示意動作映^可存在_個以上 自訂映射。例如同使用者可建立不同映射。當使用者 :啟控制面板時’其可藉由選擇示意動作映射控制然後經 由反白顯示列在自訂映射清單中移動選擇期望自訂映射。 -旦發現期望映射’使用者可簡單地反白顯示,然後實行 主要點選以便建置期望示意動作映射。之後,冑用者可改 變其設定值。 控制面板550亦可包括輸入功能選單,其亦係下拉視 窗之形式。當使用者選擇下拉視窗時,可為使用者提供一 或多個輸入功能之清單。例如,輸入功能可包括無、指 向、指向/點選、捲動、捲動及次要點選、主要點選及拖 =、次要點選及拖$、移動視窗、公開及儀錶板、網頁劉 鲁 f器操作、編輯操作、檔案操作等等。為反白顯示輸二功 月b之,使用者可具有根據使用者控制穿透清單之反白顯 示列〇 控制面板550亦可包括示意動作集區段別,其顯示與反 白顯示之輸入功能相關聯的命令,以及指派給命令之示音 動作事件。示意動作集區段僅可用於回饋,但亦可包括指 派命令給示意動作事件之功能性(反之亦然)。儘管可廣= 改變視窗,在所說明之具體實施例中,視窗可包含分離方 框,其各包含示意動作事件及與其相關聯之命令。此具體 128630.doc -45- 200847001 貝施例中,可依3x3矩陣建置方框,因此各輸入功能可存 在9個不同示意動作事件及9個不同命令。大多數情形中, 可橫跨彼此放置相反W意㈣事件。例如,可與旋 轉CW相反地定位旋轉ccw,可與收縮相反地定位擴展, I與向右擦過相反地定位向左擦過,可與^擦過相反地 疋位向上擦過0若可能, 立 此了將互補,令掃描連結至相反示 思動作事件,從而使其更直觀。 控制面板550亦可包括弦選單558。可將弦彼此疊置地展 開經由捲動列在視窗中移動。弦可包括文字識別符以 及弦指不器’其顯示如何製作弦之範例。所說明之且體實 =中,弦指示器可係手形式’其具有用於製作弦的覆蓋 手^之暗,圓圈。弦亦可包括用於指派給弦之輸入功能的 文子識別符。 一項具體實施例中,為涂罢+立&amp; 4建置不思動作映射,使用者可反 ㈣示輸入功能,然後移動至弦選單。一旦位於弦選單, ?者:糟由穿透弦清單,反白顯示期望弦然後經 (例如)實行選擇事件選擇用於該輪入功能之期望弦。之 可將輸入功能指派給弦。某些情形中,一旦使 輸入功能指派給弦,可將鈐χ汕处崎, 了將輪入功錢別符與弦-起放置在 :選:内。另彳,輸入功能可包括指示其已被指派之識別 二=在I藉由指浪之輸入功能放置勾選。操作期間, 田用者在夕點觸控表面商放置弦選單内所示之手指 時,可將手指配置辨識為弦,之後,可將與指派給弦 入功能相關聯的示意動作集設定為主要或作用中示意動作 128630.doc -46- 200847001The seven members do not have an illustrative control operation in accordance with an embodiment of the present invention. Control operations may begin at step 522, where the control is displayed: °, for example, the control object may be on-screen, dial, slider, switch, and the like. In step 524, the touch on the control object can be detected. Thereafter, at step 526, the contact configuration associated with the touch can be identified. In step 527, it may be determined whether the touch configuration matches the contact configuration associated with controlling the control object. If not, the operation may continue to step 528 where the touch is ignored. If so, the operation can continue to step 53A where the touch is monitored for a graphical action event associated with controlling the control object. If an action action event is performed, the action associated therewith can be applied to the control object to control the object to generate its command (step 43 2). In one embodiment of Figure 27, the contact configuration that initially controls the control object can be based on multiple contacts, while the ignored contact configuration can be based on a single contact. Figure 28 shows an illustrative control operation 540 in accordance with an embodiment of the present invention. You can use the control panel, for example, as shown in Figure μ, to perform control operations 128630.doc -43· 200847001 540. The operation can start at step %, where an input function is provided. The input function can have a command and a schematic action/month command to link to the command can be related to the wheeling function. For example, if the input function is edited, the command can be Cancel, Restore, Paste, Back Label, Copy, Label, Down, Cancel Restore, and Select All. The gesture action event can be an action performed by a string. Schematic action events may include rotation (clockwise and reverse time, group transitions (up, down, right, left), transitions (open, close), taps, etc. relative to each other. The action event is linked to the command. For example, 'by closing the fingers together, similar to the cut action, starting to cut the life. In addition, the complementary command (cut/paste) is usually linked to the complementary gesture event ( Open/close.) In step 544, the power can be assigned to a specific string. The string can be configured for the finger. The string can be displayed in the early months of 'for the user to highlight the specific string, 'and then select a specific Input function. The string can be assigned to multiple input functions. In step 544, the user can select the input Wei from the menu of the input function, and then select the string from the menu list of the string to assign the input function to the special string ^ For example, highlighting the input function 'and then selecting the string'. Once the string is selected, the input power month can be coupled to the string for future operation (step (10)). In some cases the 'chord menu' may include an indication of the input function assigned to it. In this particular κ &amp; example, a command has been assigned to a gesture action event. Figures 29A through G are diagrams of a control panel 550 in accordance with one embodiment of the present invention. Figure 29A shows a basic control panel, while Figures 29B through 29G show several examples of specific embodiments of the item control panel. As shown in FIG. 29A, the control panel 55A may include a schematic 128630.doc-44-200847001 action mapping control 552 in the form of a drop down window. When the user selects a drop-down window, the user can be provided with an m (four) meaning action mapping option. The default setting (usually set by the manufacturer) can be provided to the manufacturer or the user can select a custom setting value so that it can form its own gesture. There can be more than one custom mapping. For example, different mappings can be established with the user. When the user: opens the control panel, it can select the desired custom mapping by selecting the gesture mapping control and then moving through the highlight mapping list in the custom mapping list. Once the desired mapping is found, the user can simply highlight the display and then perform a primary click to construct the desired gesture mapping. After that, the user can change its setting value. Control panel 550 can also include an input function menu, which is also in the form of a drop down window. When the user selects a drop down window, the user can be provided with a list of one or more input functions. For example, input functions can include no, pointing, pointing/clicking, scrolling, scrolling and sub-point selection, main point selection and dragging, sub-point selection and dragging $, mobile window, public and dashboard, web page Liu Lu f device operations, editing operations, file operations, and so on. In order to display the binary power b, the user may have a reverse display display according to the user control penetration list. The control panel 550 may also include a gesture set section, and the input function of the display and the reverse display. Associated commands, as well as voice action events assigned to the command. The gesture set is only available for feedback, but may also include the assignment of commands to the functionality of the gesture (and vice versa). Although the window can be varied widely, in the particular embodiment illustrated, the window can include separate frames, each of which includes a gesture event and a command associated therewith. This specific 128630.doc -45- 200847001 shell example, the box can be built according to the 3x3 matrix, so each input function can have 9 different schematic action events and 9 different commands. In most cases, the opposite (four) events can be placed across each other. For example, the rotation ccw can be positioned opposite to the rotation CW, and the extension can be positioned opposite to the contraction. I can be wiped to the left by rubbing the opposite direction to the right, and can be rubbed up by 0 in the opposite direction to the wiper. Complementary, making the scan link to the opposite of the action action event, making it more intuitive. Control panel 550 can also include a string menu 558. The strings can be spread over each other and moved through the window via the scrolling column. The string may include a textual identifier and an example of how the string is displayed. Illustrated and physically = medium, the chord indicator can be in the form of a hand&apos; which has a dark, circular circle for the covering hand of the chord. The string may also include a text identifier for assigning an input function to the string. In one embodiment, the mapping of the motion map is applied to the coating and the user can reverse (4) the input function and then move to the string menu. Once in the string menu, the player: the list of penetrating strings, highlights the desired string and then selects the desired string for the wheeling function via, for example, a selection event. The input function can be assigned to the string. In some cases, once the input function is assigned to the string, you can place it in the line: Select: within the selection. Alternatively, the input function may include an indication that it has been assigned. 2 = Placed in the I input by the finger input function. During operation, the field user can recognize the finger configuration as a string when the touch surface operator places the finger shown in the string menu, and then the set of gestures associated with the function assigned to the string entry function can be set as the main Or the action in action 128630.doc -46- 200847001

圖29B顯示一範例,其中可開啟輸入功能選單,並且使 用者可指派公開儀錶板輸入功能給4手指弦。另外,厂立 動作集可顯示與公開/儀錶板輸入功能相關聯之命令,2 及用於起始命令之示意動作事件。例如,向上擦過可起始 公開全部,向右擦過可起始公開桌面,向左擦過可起始^ 開應用程式,向下擦過可起始儀錶板。Figure 29B shows an example in which the input function menu can be turned on and the user can assign a public dashboard input function to the 4-finger string. In addition, the factory action set displays the commands associated with the public/dashboard input function, 2 and the schematic action events used to initiate the command. For example, swipe up to start publicizing all, wipe to the right to start the open desktop, wipe to the left to start the application, and swipe down to start the dashboard.

圖29C顯示一範例,其中可開啟輸入功能選單,並且使 用者可指派主要點選及拖矣輸入功能給3手指弦。另外, 示意動作集可顯示與點選及拖良輸入功能相關聯之命令, 以及用於起始命令之示意動作事件。例如,任何方向上之 方向移動可用於拖良物件,並可使用敲擊以起始點選。 圖29D顯示一範例,其中可開啟輸入功能選單,並且使 用者可指派檔案操作輸入功能給拇指+1手指弦。另外,示 意動作集可顯示與檔案操作輸入功能相關聯之命令,以及 用於起始命令之示意動作事件。例如,旋轉擦過ccw可用 於起始開啟,旋轉擦過CW可用於起始關閉,擴展可用於 新增,收縮可用於保存。 圖29E顯示一範例,其中可開啟輸入功能選單,並且使 用者可指派編輯操作輸入功能給拇指+2手指弦。另外,示 〜動作集可顯示與編輯操作輸入功能相關聯之命令,以及 :於起始命令之示意動作事件。例如’向上擦過可起始復 程式向右擦過可起始標籤,向左擦過可起始後退標籤應用 向下擦過可起始取消復原。另外,旋轉擦過ccw可 128630.doc -47- 200847001 起始取 &gt;肖’旋轉擦過c 上,收縮可起始剪下二部’擴展可起始貼 圖29叫f ~ 料,敲擊可起始複製。 用者可指派網頁劉覽哭摔1: 功能選單,並且使 外,示音動作I 功能給梅指+3手指弦。另 之命令:以及木可顯示與網頁劉覽器操作輸入功能相關聯 捧二二起始命令之示意動作事件。例如,向上Fig. 29C shows an example in which an input function menu can be turned on, and the user can assign a main point selection and drag input function to the 3 finger string. In addition, the illustrated action set can display commands associated with the click and drag input functions, as well as gesture action events for the start command. For example, moving in any direction can be used to drag objects, and taps can be used to initiate a click. Figure 29D shows an example in which the input function menu can be turned on and the user can assign a file operation input function to the thumb +1 finger string. In addition, the set of action actions can display commands associated with the file operation input function, as well as gesture action events for the start command. For example, a rotary wipe ccw can be used to initiate the start, a rotary wipe over CW can be used to initiate the close, an extension can be used for the new, and a shrink can be used for the save. Figure 29E shows an example in which the input function menu can be turned on and the user can assign an edit operation input function to the thumb + 2 finger string. In addition, the action-action set can display commands associated with the edit operation input function, as well as: a schematic action event at the start command. For example, 'Swipe up to start the program to the right to wipe the start label, wipe to the left to start the back label application. Press down to start canceling the restore. In addition, the rotating wipe ccw can be 128630.doc -47- 200847001 Start to take > Xiao's rotation to rub on c, shrink can start to cut two 'extension can start map 29 called f ~ material, tap can start copy. The user can assign a web page to Liu Ming Cry 1: function menu, and make the sound action I function to the plum finger +3 finger string. Another command: and wood can display the action action event associated with the web browser operation input function. For example, up

後退…下t止’向右擦過可起始向前,向左擦過可起始 過可起始重新整理…卜,向左上擦過可起 尋°找二擦過可起始聚光光線,向左下擦過可起始 寸找,向右下擦過可起始書簽。 用:員下乾例,其中可開啟輸入功能選單,並且使 Μ曰派指向輸入功能給拇指+4手指弦。另外,示咅動 2可顯示與指向輸入功能相關聯之命令,以及用於㈣ 示意動作事件。例如,任何方向上之方向移動可用 圖30顯示根據本發明之一項具體實施例的說明性控制操 ^ 580。可使用控制面板’例如圖29所示者,實行控制操 作—操作可始於步驟582 ’在此步驟提供輸人功能之清 早。步驟584中,可提供與輸入功能相關聯之命令清單。 步驟靖,可提供弦清單。步驟谓中,可提供與弦相關 聯之不意動作事件清單。步驟別中,可將輸人功能指派 給弦(反之亦然)。步驟592中,可將命令指派給示意動作事 件。替代操作中,操作可包括提供所有可用命令之一 清單,而非使清單與特定輸人功能相Μ。使用者可指派 128630.doc -48- 200847001 任何命令至輸入功能。同樣在替代操作中,操作可包括接 相關聯。使用者可而非使清單與特定弦 用#了才曰派不意動作事件給弦。 /考圖29冑用者可藉由從輸人功能之選單選擇幹 能,然後從弦之選單、眚簞、登 雨 早α早k擇弦,將輸入功能指派給特定 列如反白顯示輸入功能,然後選擇弦)。一旦選擇 :’可將輸人功能聯結至該弦以供未來操作。某青Retreat...Under the end of 'Through to the right to start forward, to the left to rub can start, can start to reorganize... Bu, rub it up to the left to find it. Look for the second wipe to start the spotlight, and wipe it to the left. You can find the initial inch and wipe it to the lower right to start the bookmark. Use: The staff can do the following, which can open the input function menu and make the Μ曰 指向 point to the input function to the thumb + 4 finger string. In addition, the trigger 2 can display commands associated with pointing to the input function and for (iv) signaling action events. For example, directional movement in any direction may be used in Figure 30 to illustrate an illustrative control operation 580 in accordance with an embodiment of the present invention. The control operation can be performed using a control panel&apos; such as that shown in Fig. 29 - the operation can begin at step 582' where the input function is provided early. In step 584, a list of commands associated with the input function can be provided. Step Jing, can provide a list of strings. In the step-by-step, a list of unintentional action events associated with the chord can be provided. In the step, the input function can be assigned to the string (and vice versa). In step 592, the command can be assigned to the gesture action event. In an alternative operation, the operation may include providing a list of all available commands rather than making the list contrary to a particular input function. The user can assign 128630.doc -48- 200847001 any command to the input function. Also in an alternative operation, the operations may include an association. Instead of making the list and the specific chords use #, the user can send an unintentional action event to the string. / Test Figure 29 users can select the dry energy from the menu of the input function, and then select the input function from the menu of the string, the 眚箪, the rain, and the input function to the specific column, such as the reverse display input. Function, then select the string). Once selected: 'The input function can be linked to the string for future operation. Certain green

中,弦選單可包括指派給其之輸人功能的指示。另^使 :者可猎由在不意動作上移動命令指派命令給示意動作事 如’使用者可選擇並拖$特定命令,並在期望示意 件上將其放下。之後,可將命令聯結至該示意動作 事件以供未來操作。 圖“、員不根據本發明之一項具體實施例的說明性示意動 乍映射方法600。映射_可始於步驟6〇2,在此步驟將系 統放置於示意動作咏 卞映射杈式中。一旦處於示意動作映射模 可在γ驟604内偵測觸控。之後,步驟6〇6中,可儲存 與觸控相關聯之接觸配置。例如,可產生接觸配置之影 像:之後,步驟608中,可針對示意動作事件監視觸控。 後乂驟61 〇中’可儲存示意動作事件。之後,步驟612 中立可將叩令指派給接觸配置/示意動作事件。一旦離開 〜動作映射杈式’可藉由放下接觸配置以便可加以辨 識,然後實行示意動作事件來起始命令。 圖32顯不根據本發明之一項具體實施例的說明性示意動 卞映射方去620。不意動作映射方法62〇可始於步驟622, 128630.doc -49- 200847001 在此步驟可開啟一命令選單。命令選單可為在應用程式標 通中找到的該等選單之任何一個(例如標案、編輯、檢 視、***、格式、工具等)。當開啟命令選單時,可顯示 咋7之用單(步驟624)。步驟624後,可經由使用者控制在 p 7 β單中的命令上移動反白顯示列。之後,步驟626 ttr 可在反白顯示一命令時彳貞測觸控表面上之觸碰。之 後,步驟628中,可成像及儲存與觸控相關聯之接觸配 置。之後,步驟630中,可針對示意動作事件監視觸控。 之後,步驟632中,可在實行時成像及儲存示意動作事 件。之後,步驟634中,可將接觸配置及示意動作事件指 派給反白顯示之命令。一旦離開示意動作映射模式,可藉 由放下接觸配置以便可加以辨識,然後實行示意動作事件 來起始命令。 圖33顯示根據本發明之一項具體實施例的說明性示意動 作回饋方法640。操作640可始於步驟642,在此步驟偵測 觸控。之後,在步驟644中,可辨識接觸配置。之後,步 驟646中,可至少部分根據接觸配置載入示意動作集。2 後,步驟648中,可在觸控期間偵測暫停(例如在一時間週 =内不實行示意動作事件之後,步驟㈣中,可顯示示 似動作集,以便使用者可檢視命令及與其聯結之示意動作 事件(例如參見圖34)。之後,步驟652中,當示意動作 始時或預設逾時後可從顯示器移除示意動作集。 圖34顯示根據本發明之一項具體實施例的說明性示 作檢視視窗_。示意動作檢視視窗可為_示意動:選 128630.doc •50· 200847001The string menu may include an indication of the input function assigned to it. In addition, the player can hunt by assigning a command to the gesture action on the unintentional action, such as 'the user can select and drag the $ specific command, and drop it on the desired gesture. The command can then be linked to the gesture event for future operation. The diagram "is not an illustrative schematic mapping method 600 in accordance with an embodiment of the present invention. The mapping_ may begin in step 6〇2, where the system is placed in a schematic action mapping scheme. Once in the gesture mapping mode, the touch can be detected in the gamma step 604. Thereafter, in step 6〇6, the contact configuration associated with the touch can be stored. For example, an image of the contact configuration can be generated: after that, in step 608 The touch can be monitored for the action event. After the step 61, the action event can be stored. After that, the step 612 can assign the command to the contact configuration/schematic action event. Once it leaves the action map, it can be The command is initiated by dropping the contact configuration so that it can be recognized, and then performing a gesture action event. Figure 32 shows an illustrative schematic mapping of the 620 according to an embodiment of the present invention. Beginning at step 622, 128630.doc -49- 200847001 A command menu can be opened in this step. The command menu can be any of the menus found in the application gating. For example, standard, edit, view, insert, format, tool, etc.) When the command menu is opened, the menu for 咋7 can be displayed (step 624). After step 624, it can be controlled by the user in the p 7 β list. The command reverses the display column. After that, step 626 ttr can detect the touch on the touch surface when a command is displayed in reverse video. Then, in step 628, the contact configuration associated with the touch can be imaged and stored. Thereafter, in step 630, the touch can be monitored for the gesture event. Thereafter, in step 632, the gesture event can be imaged and stored during execution. Thereafter, in step 634, the contact configuration and the gesture event can be assigned to the inverse The white display command. Once the gesture mapping mode is removed, the command can be recognized by dropping the contact configuration and then performing a gesture action event. Figure 33 shows an illustrative schematic action in accordance with an embodiment of the present invention. Feedback method 640. Operation 640 can begin at step 642, where touch is detected. Thereafter, in step 644, the contact configuration can be identified. Thereafter, in step 646 The gesture set can be loaded at least in part according to the contact configuration. 2 After step 648, the pause can be detected during the touch (for example, after the gesture event is not performed within one time period =, in step (4), the display can be displayed. The set of actions so that the user can view the command and the gesture action event associated with it (see, for example, Figure 34). Thereafter, in step 652, the set of gestures can be removed from the display when the gesture begins or after a preset timeout. 34 shows an illustrative view window _ according to a specific embodiment of the present invention. The schematic action view window can be _ gesture: select 128630.doc • 50· 200847001

早,其彈出關於如何實行示意動作事件及若實行示意動作 事件應起純種命令的包含資訊。制者因此可在實行示 意動作事件前快速將示意動作㈣印至記憶體^示意動 作視窗_可係半透明或不透明視窗,其係顯示於第-觸 碰點(或暫停)附近以及任何其他顯示之資訊上。儘管可廣 泛改變視窗,在所說明之具體實施例中,4見窗660可包含 分離方框662’其各包含示意動作事件咖及與其相關聯之 命令667。此具體實施例中,可依3χ3矩陣建置方框,因此 可存在9個不同示意動作事件及9個不同命令。大多數情形 中,可橫跨彼此放置相反實行之示意動作事件。例如,可 與旋轉cw相反地定位旋轉ccw,可與收縮相反地定位擴 展’:與向右擦過相反較位向左擦過,可與向下擦過相 _ $位向上擦。^彳能,可將互補命令掃描連結至相 反丁心動作事件,從而使其更直觀。例如,在剪下及貼上 j下可與收縮相關聯,而貼上可與擴展相關聯。 另外另一個可與敲擊相關聯(以圓圈顯示)。 圖35顯示根據本發明之一項具體實施例經由觸控螢幕實 鉍的忒明性鍵盤方法68〇。方法68〇可始於步驟Μ:,在此 ’: 動觸控螢幕鍵盤(例如顯示並準備接收輸入)。之 後,步驟684中,可針對按鍵輸入監視觸控表面(例如在顯 不:按鍵上敲擊的手指)。同時,步驟686中,可針對擦過 不思動作事件監視觸控表面。若擦過示意動作事件係向右 擦過,可實行空格命令(步驟688及69〇)β若擦過示意動作 事件係向左擦過,可實行退格/刪除命令(步驟692及694)。 128630.doc -51 - 200847001 若擦過示意動作事件係向τ擦過,可實行輸人/返回命令 Ο驟696及698)。右擦過示意動作事件係向上擦過,可實 行分頁命令(步驟700及7〇2)。然而,應注意,方法不限於 該等命令,可實行其他命令。 人由於要評估之弦、示意動作事件、命令有如此多的組 有幫助的可係隔離優勢及弱點,並發展哪些弦及示意 動料件與哪些命令較佳配合的引導理論。—般咸信藉由 匹配取佳手指組合與最頻繁命令最大化長期使用者體驗。 f通手指組合僅應用於較不頻繁命令 '然而,由於細微問 題’例如串擾,只遵循此原則不足以建立優秀使用者體 驗。 2顯示根據本發明之—項具體實施例的說明性示意動 作設計程序72〇。設計程序72〇包括步驟Μ],在此步驟可 額定或分級最容易弦。例如,弦可根據速度、可靠性、舒 適度、及使用容易性、簡單性等加以額定。此亦可基於觸 控表面之尺寸,因為較小表面無法容納所有手指組合。圖 37顯不用於額定弦之矩陣的一範例。應明白,此矩陣係範 例形式,而非限制。 設計程序亦包括步驟7〇4,在此步驟可額定或分級操作 頻率。例如,特定使用者或使用者群組實行特定操作集之 頻繁度。頻率分級可依據操作相對於彼此之比例(例如網 頁瀏覽對編輯)變更。圖38顯示用於分級一些操作之操作 頻率的矩陣之一範例。應明白,此清單係範例形式而非限 制’因為可存在更多操作。 义 128630.doc -52- 200847001 設計程序亦可包括步驟鳩 弦分組或盥最锢够/外只貝上將取容易 最不頻繁操:=作匹配,而實質上將最難㈣ 因Γ==括精細調諧群組。精細調譜可基於許多 明白,直觀性、串擾、可感測組織等。應 又口 t耘序可能不像匹配最容易弦 簡單。某些情形中,hh 取呆作-樣 義。同樣,= 可能對於最頻繁操作無意 7 4呆作可偏移,同時仍嘗試保持容易盥最噸 繁以及困難與最不頻繁。 易/、取頻 圖39顯示根據本發一 作740。&quot;】握你 $具心施例的說明性控制操 動作映射始於步驟742,在此步驟提供基礎示意 m甘 土楚不思動作映射可包括示意動作偏好設定, 弦、命令示意動作事件等。步驟742後,操作 為、驟744,在此步驟第—程式作 可為文字虛棟名。j J X 式 处里耘式、媒體管理程式、試曾 輯程式、綠製程式等。步驟746中,可:第二媒體編 同時根據第-程式調整基礎示意動:式作用的 指示基礎示咅動#地&amp; ^例如,程式可 楚丁似動作映射以改變弦及示意 程式之意義。步驟748中,第二程 ^於此特疋 可在篦-妒m J作用。步驟750中, 映射式作用的同時根據第二程式調整基礎示意動作 射圖3- 媒體編輯程式可如下調整示意動作^ 物件二 兩個手指用於移動、縮放、旋轉 手“於主要拖,;四個手指用於捲動平移. 128630.doc -53- 200847001 5個手指用於如&amp; 心 向,同時文字處理程式可如下镅敕 作映射·· 2個手和爾认拉a 士 a 卜調整不意動 ㈤予#日肖於指向文字游標 羯 字,4個丰鈔田#、 w卞鈿用於選擇 子個手^曰用於分頁、開始線及結束線。 擇文 圖40顯示根據本發 作⑽。控制摔作可虫Λ動择具β實施例的說明性控制操 指弦實施的二作=關聯’例如採用三手Early, it pops up information about how to perform a gesture action event and if a gesture action event is to be performed. The controller can therefore quickly print the gesture (4) to the memory ^ gesture window _ can be a translucent or opaque window, which is displayed near the first touch point (or pause) and any other display before the gesture event is executed. Information. Although the window can be widely changed, in the illustrated embodiment, window 660 can include separate blocks 662' each containing a gesture action event and a command 667 associated therewith. In this embodiment, the blocks can be constructed in accordance with the 3χ3 matrix, so there can be 9 different gesture events and 9 different commands. In most cases, a schematic action event that is reversed can be placed across each other. For example, the rotation ccw can be positioned opposite to the rotation cw, and the extension can be positioned opposite to the contraction: the opposite side is wiped to the left by rubbing it to the right, and can be rubbed up with the _$ bit. ^彳, can be combined to scan the complementary command to the opposite event, making it more intuitive. For example, cut and paste j can be associated with shrinkage, and paste can be associated with extensions. The other can be associated with a tap (shown in a circle). Figure 35 shows a illustrative keyboard method 68 经由 via a touch screen in accordance with an embodiment of the present invention. Method 68 can begin at step Μ:, here:: Touch the touchscreen keyboard (eg, display and prepare to receive input). Thereafter, in step 684, the touch surface can be monitored for key input (e.g., on the display: the finger tapped on the button). At the same time, in step 686, the touch surface can be monitored for the wipe action event. If the wiped action event is swiped to the right, a space command can be executed (steps 688 and 69). β If the action is swiped, the event is swiped to the left, and the backspace/delete command can be executed (steps 692 and 694). 128630.doc -51 - 200847001 If the action is swiped to τ, the input/return command can be executed (steps 696 and 698). The right wipe action signal is swiped upwards, and the page break command can be executed (steps 700 and 7〇2). However, it should be noted that the method is not limited to these commands and other commands may be implemented. Because of the many strings that need to be evaluated, the action events, and the command, there are so many groups that can help isolate the strengths and weaknesses, and develop the guiding theory of which strings and which commands are better matched with which commands. As always, the salty letter maximizes the long-term user experience by matching the best finger combination with the most frequent commands. The f-pass finger combination is only applied to less frequent commands. 'However, due to subtle issues such as crosstalk, it is not sufficient to follow this principle to establish an excellent user experience. 2 shows an illustrative schematic design procedure 72 of a particular embodiment of the present invention. The design program 72 includes a step Μ] in which the most easily chord can be rated or graded. For example, strings can be rated based on speed, reliability, comfort, ease of use, simplicity, and the like. This can also be based on the size of the touch surface because the smaller surface cannot accommodate all finger combinations. Figure 37 shows an example of a matrix of nominal chords. It should be understood that this matrix is a generic form, not a limitation. The design procedure also includes steps 7〇4, where the frequency can be rated or graded. For example, the frequency with which a particular user or group of users implements a particular set of operations. Frequency grading can vary depending on the ratio of operations relative to each other (e. g., web browsing for editing). Figure 38 shows an example of a matrix for grading the operating frequencies of some operations. It should be understood that this list is an example rather than a limitation 'because there may be more operations. Yi 128630.doc -52- 200847001 The design program can also include the steps of the chord grouping or 盥 盥 / 外 外 只 只 只 上 上 上 上 上 = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = Includes a fine tuning group. Fine tuning can be based on a lot of understanding, intuitiveness, crosstalk, sensible organization, and more. It should not be the same as matching the easiest string. In some cases, hh takes a dignified-like meaning. Similarly, = may be unintentional for the most frequent operations, while still trying to keep the most ton, difficult and least frequent. Easy/Frequency Acquisition Figure 39 shows a 740 according to the present invention. &quot;] The illustrative control action map that holds your $intentional example begins at step 742, where the basics are provided to indicate that the action mapping can include gestures, chords, commands, action events, etc. . After step 742, the operation is step 744. In this step, the program can be a virtual name. j J X type, media management program, trial program, green program, etc. In step 746, the second media can be based on the first program to adjust the basis of the motion: the indication of the role of the action indicates that the grounding &amp; ^, for example, the program can be similar to the action mapping to change the meaning of the string and the schematic program . In step 748, the second pass ^ can be used in 篦-妒m J. In step 750, the mapping function is performed according to the second program, and the basic gesture is taken according to the second program. The media editing program can adjust the gestures as follows. ^ Objects Two fingers are used to move, zoom, rotate the hand "on the main drag,; four Fingers are used for scrolling translation. 128630.doc -53- 200847001 5 fingers are used for &amp; heart direction, and the word processing program can be mapped as follows. · 2 hands and er er pull a 士 a 卜 adjustment Do not move (five) to #日肖 in the text cursor 羯 word, 4 Feng Shou Tian #, w卞钿 is used to select the child hand ^ 曰 for pagination, start line and end line. Select text 40 shows according to this episode (10). Controlling the fall of the worm's swaying option. The illustrative control of the embodiment of the finger is performed by the second function = association 'for example, using three hands

驟決定是否價測到向右擦過。若是,摔/可=在此步 ⑹,在此步驟起始第一公開功能。若否;;至步驟 :驟6在此步驟決定是否谓測到向左擦過。若是二 /至步驟768,在此步驟起始第二公 + 否,操作可繼續至步驟77〇, b右 上柊ift Μ曰 隹此步驟決定是否偵測到向 上裇過。右疋,操作可繼續 -八叫丄 乂驟772,在此步驟起始第 二Α開功能。若否,操作可 定是否债測到向下捧過。若β 7 ,在此步驟決 一 右疋,刼作可繼續至步驟776, 在匕4起始紐板功^若^,操作可返回762。 圖:顯不根據本發明之一項具體實施例的說明性控制操 作780。控制操作可始於 ^ ^ 哪782,在此步驟偵測基弦。例 基弦可為三手指弦。觸批&gt; 觸控仃程期間,偵測-暫停(步 驟784)。之後,步驟786中, —Λ Γ因為基弦之暫停開啟一内 谷相關選單。之後,在步驟 &quot; 中’決定是否镇測向上或 向下滑動。若是,操作可繼螬 、’至步驟790,在此步驟根據 •'月動運動在内容相關選單中移動反白顯示列。之後,在步 驟792中,決定是否偵測 〗詞岐擊。若是,操作可繼續至步 在此步驟實施反白顯示之項目的功能性。 128630.doc -54- 200847001 儀錶板係界面工具集之首頁:使您實行共同任務並為您 提供對資訊之快速存取的微型應用程式。採用擦過,儀錶 板可連同界面工具集出現,其立即將使用者帶入資訊世 界:即時天氣、證券收報機、航班資訊及更多。儀錶板可 同樣容易地消失,以便使用者回到先前之工作。儀錶板由 加州 Cupertino的 Apple lnc·製造。 公開使使用者可採用單一擦過以及無法模仿的傾倒風度 立即存取任何開啟視窗。縮圖顯示所有開啟視窗,檢視當 絀應用程式視窗或隱藏所有視窗以快速定位桌面上之檔 案。公開由加州Cupertino的Apple Inc.製造。 圖42至44顯示根據本發明之多個具體實施例的各種示意 動作映射。各圖式包括四個標題:1}手,2)手指數目 (弦),3)示意動作事件,及4)與示意動作事件相關聯之行 為或動作。 圖45至48顯示根據本發明之數個具體實施例的各種示意 動作集。所有料圖式巾,示意動作集包括由箭頭顯示之 各種動作(擦過、旋轉、擴展/收縮),以及由圓圈顯示之一It is decided whether the price is detected to the right. If yes, fall/can = at this step (6), the first public function is started at this step. If no;; to step: Step 6 In this step, it is determined whether it is detected that it has been wiped to the left. If it is 2 / to step 768, at this step start the second public + no, the operation can continue to step 77, b right up ift Μ曰 隹 步骤 步骤 隹 隹 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 Right 疋, the operation can continue - 八叫 乂 772 772, in this step start the second open function. If not, the operation can determine whether the debt is measured down. If β 7 , right-click on this step, the operation can continue to step 776, and the operation can return to 762 at the start of the new board. Figure: An illustrative control operation 780 that is not in accordance with an embodiment of the present invention. Control operations can begin at ^ ^ 782, where the basestring is detected. The base string can be a three-finger string. During the touch batch &gt; touch-time, detect-pause (step 784). Thereafter, in step 786, - Λ 开启 opens a valley related menu because of the pause of the base chord. After that, in step &quot; in the 'decision' whether the town test slides up or down. If so, the operation can be followed by ’, to step 790, in which the column is displayed in the content related menu according to the 'monthly motion'. Thereafter, in step 792, it is determined whether to detect the word sniper. If yes, the operation can continue to step. In this step, the functionality of the highlighted item is implemented. 128630.doc -54- 200847001 The Home of the Dashboard Interface Tool Set: A micro-app that enables you to perform common tasks and provide you with quick access to information. With wiping, the dashboard can come with an interface tool set that immediately brings users to the information world: instant weather, securities tickers, flight information and more. The dashboard can disappear as easily as the user returns to the previous job. The dashboard is manufactured by Apple Lnc. of Cupertino, California. The disclosure allows the user to instantly access any open window with a single wipe and a dumping style that cannot be mimicked. The thumbnail shows all open windows, view the application window or hide all windows to quickly locate the file on the desktop. Published by Apple Inc. of Cupertino, California. Figures 42 through 44 show various schematic motion maps in accordance with various embodiments of the present invention. Each schema includes four headings: 1} hand, 2) number of fingers (chord), 3) gesture action events, and 4) behavior or action associated with the gesture event. Figures 45 through 48 show various illustrative sets of actions in accordance with several embodiments of the present invention. All pattern drawings, the set of action actions include various actions (wipe, rotate, expand/contract) displayed by arrows, and one of the circles displayed

或多個敲擊。其亦可包括聯結至所屬運動之命令。可根據 使用者而要將示,⑤動作集應用於不同㊣。—範例總,可將 圖45之示意動作集應用於拇指+1手指弦,可將圖牝之示意 動作集應用於拇指+2手指弦,可將圖47之示意動作集應用 ;拇私+3手扎弦’以及可將圖48之示意動作集應用於伸展 拇才曰及3手指弦。應瞭解,示意動作集及弦係藉由範例方 式而非限制。例如’在示意動作集情形中,可使用不同命 128630.doc -55- 200847001 々,可使用不同示思動作事件,可使用不同指派等等。 現在將說明關於圖8至11之額外具體實施例。 圖49顯示根據本發明之一項具體實施例的說明性示咅動 作操作800。射800可始於步驟8〇2,在此㈣僅^第 一手指。之後,在步驟8〇4,可決定第一手指之狀熊。 如’狀態可對應於手指之特定方位、手指相對於觸控表= 之位置、手指係移動還是固定、手指動作之方向等。Or multiple taps. It may also include an order to connect to the sport to which it belongs. According to the user, the 5 action sets can be applied to different positives. - For example, the schematic action set of Figure 45 can be applied to the thumb +1 finger string, and the schematic action set of the figure can be applied to the thumb + 2 finger string, and the schematic action set of Figure 47 can be applied; The hand string is 'and the schematic action set of Figure 48 can be applied to the stretched thumb and 3 finger chords. It should be understood that the illustrated action sets and strings are by way of example and not limitation. For example, in the case of a schematic action set, different lives 128630.doc -55- 200847001 can be used, different cue action events can be used, different assignments can be used, and so on. Additional specific embodiments with respect to Figures 8 through 11 will now be described. Figure 49 shows an illustrative operation 800 in accordance with an embodiment of the present invention. Shooting 800 can begin at step 8〇2, where (4) only the first finger. Thereafter, in step 8〇4, the bear of the first finger can be determined. For example, the state may correspond to a specific orientation of the finger, the position of the finger relative to the touch table =, the movement or fixation of the finger system, the direction of the finger movement, and the like.

後,步驟_中’可债測一或多個額外手指。例二、,二 一手指外,可❹j不同第二手指。步驟8〇8中,可決^ 一 或多個額外手指之狀態。之後,步職0中,可根據= 及額外手指相對於彼此之狀態時序實施不同輪i式 同輸入模式可廣泛變化’並且可係選自本文所述 容或任何適當或已知輸人模式。輸人模式之— ㈣式或拖夷模式。其他模式包括,例如,捲動、平= 專0 一手指之狀態可包括固定或移動,第二丰 指之狀態可包括第二手指存 一手 刭笛一丰扣仔隹一否此乾例中,若僅偵測 到弟η曰,可實施指向操作。$外 / 定時偵㈣帛二手彳旨,可 手指係固 /筐一车扣# 、扣向刼作切換至拖曳操作,若 在弟-手扣移動時偵測到第二 此做,使用者 向細作。如 如,#用去τ* 矣操作間來回切換。例 使用者可在觸控表面上移動一個手指 資料爽上定位游標(例如 :檔案 或游標,將第二手指放下至f 4夹上停止手指 至觸控表面上,從而指示期望拖 128630.doc •56- 200847001 ,資料夹’然後將第-手指移動至另-位置以拖,檔 ==至新位置。使用者可保持移動其手指,從而四處 移動檔木資料夹,或簡單地抬起第—手指以最終定位營幕 内的槽案資料夾。 圖50顯示根據本發明之一 H,、體貝轭例的說明性示意動 作才呆作820。操作820可始於步驟82 一 於步驟822,在此步驟僅偵測第 2 :之後,在步驟824中,可決定第一手指之狀態。 Γ 對應於手指之特定方位'手指相對於觸控表 之位置、手指係移動還是固定、手指動作之方向等 後,步驟826中,可名筮 主此占 主收者 了在弟一手指處於第一狀態時採用第一 崎人模式。步驟828中,可監視修改符手指事 &lt; °可精由除第—手指外的手指提供修改符手指事件。例 如其可為弟二手指。事件可為望-主杜. 醫了為弟-手指之存在或其可為 二類i之不意動作事件或特定弦。之後,步驟83〇中, j 一手“處於第二狀態且偵測到修改符手指事件時可實 施第二模式。例如’修改符事件可為靠近第—手指之第2Thereafter, step _ middle may measure one or more additional fingers. Example 2, two fingers, can be different from the second finger. In step 8〇8, the state of one or more additional fingers can be determined. Thereafter, in Step 0, different rounds of the same input mode can be widely varied depending on the state timing of the = and additional fingers relative to each other' and can be selected from the description herein or any suitable or known input mode. Input mode - (4) or drag mode. Other modes include, for example, scrolling, flat=specific 0, the state of one finger may include fixed or moving, and the state of the second rich finger may include the second finger storing one hand, the flute and the other, in this case, If only the younger η曰 is detected, the pointing operation can be performed. $ outside / timing detection (four) 帛 second-hand purpose, can be finger-fastened / basket one car buckle #, buckle to switch to drag operation, if the second-hand is detected when the brother-handle moves, the user Fine work. For example, #Use the τ* 矣 operation to switch back and forth. For example, the user can move a finger on the touch surface to locate the cursor (for example: file or cursor, put the second finger down to the f 4 clip and stop the finger to the touch surface, thereby indicating the desired drag 128630.doc • 56- 200847001, folder 'then move the first-finger to another position to drag, file== to the new position. The user can keep moving his finger to move the wood folder around, or simply lift the first- The finger is positioned to ultimately locate the slot file in the camp. Figure 50 shows an illustrative schematic action of the body yoke example in accordance with one of the present invention H. Operation 820 can begin at step 82 and at step 822. This step only detects the second: after, in step 824, the state of the first finger can be determined. 对应 Corresponding to the specific orientation of the finger, the position of the finger relative to the touch table, the movement of the finger system or the fixed direction of the finger movement After the step 826, the name of the master can be used as the primary receiver. When the first finger of the brother is in the first state, the first satin mode is adopted. In step 828, the modifier can be monitored and the target can be removed. First hand The outer finger provides a modifier finger event. For example, it can be a second finger of the brother. The event can be a dream-main Du. He is a brother--the existence of a finger or it can be a second-class i unintentional action event or a specific string. In step 83, j may perform the second mode when in the second state and the modifier finger event is detected. For example, the modifier event may be the second near the first finger.

手指的存在。—範例中,第—模式係指向模式,第二模Z 係拖复模式。當然,可根據需要將具體實施例應用於其他 模式。 、 圖51顯示根據本發明之一項具體實施例的說明性示意動 =作840。操作84〇可始於步驟m,在此步驟最初债測 弟:手指。之後,步驟844中,可在移動第-手指時實施 ’、第+ *曰4目關聯之第一模式。&lt; 後’步驟州中,可在 偵測第手私後偵測第二手指。步驟848中,#第—手指 128630.doc •57· 200847001 二=_到第二手指之存在時可將第—模式切換至第 位置:弟二手指之存在可為關於(例如)觸控面板之特定 =的特定位置,或者相對於第—手指之特定位置。一範 中,可ΓΓ接近第一手指之第二手指。之後,步驟85。 〒 了在移動第一手指睹i姿田β 式。-範财m可仲手指實施第二模 髮模戎。、式可#、彳日向模式’第二模式可係拖 式。“田可根據需要將具體實施例應用於其他模The presence of a finger. - In the example, the first mode is the pointing mode and the second mode is the drag mode. Of course, the specific embodiment can be applied to other modes as needed. Figure 51 shows an illustrative schematic = 840 in accordance with an embodiment of the present invention. Operation 84〇 can begin at step m, where the initial debt tester: finger. Thereafter, in step 844, the first mode associated with the ', + + 4' target can be implemented when the first finger is moved. &lt; After the step state, the second finger can be detected after detecting the first hand. In step 848, #第—finger 128630.doc •57· 200847001 two=_ can switch the first mode to the first position when the second finger is present: the presence of the second finger can be related to, for example, a touch panel The specific location of a particular = or relative to the specific location of the first finger. In one example, the second finger of the first finger can be approached. After that, step 85. 〒 I moved the first finger 睹i 姿田 β. - Fan Caim can implement the second model in the middle of the finger. , the formula can be #, the day mode. The second mode can be dragged. “Tian can apply specific embodiments to other models as needed.

作=:根Λ本發明之一項具體實施例的說明性示意動 =丰。#作_可始於步_,在此步驟決定是否 :測=指。根據系統需要,可識別或不識別單一手 才曰。例如,某些情形中, 作,而其他情形中,若心疋手指則僅可實施操 法。若#龍⑽ #相任何單-手指則可實施該方 驟決定是否移動手:若:::Γ至步驟864,在此* 向操作可結束。若不移Π不牛再:測到第-手指,指 -手指固料是㈣測第 =了决疋在弟 如。若是,操插作可返回步驟 第二手指後是否移動第:手:8;°广步驟決定在㈣ -手指-起移動。大多J::可或不要求第二手指與第 不再加以摘測。若 月厂,可抬起第二手指,以便 _。若移動第-手指=—手指,則操作可返回步驟 根據第-手指心行=繼續至步驟心在此步驟 曳操作。若不再偵測到第一手 128630.doc -58- 200847001??? =: An illustrative schematic of a particular embodiment of the invention. #作_ can start at step _, in this step decide whether: test = refers. A single hand can be identified or not identified depending on the needs of the system. For example, in some cases, it is done, and in other cases, if the finger is licking, only the operation can be performed. If #龙(10)# is any single-finger, you can implement this method to decide whether to move the hand: If :::: Γ to step 864, the operation can end here. If you don't move, don't move the cow again: the first finger is measured, the finger solid is (four) measured = the final is in the younger. If yes, the operation can be returned to the step after the second finger is moved: hand: 8; ° wide step is determined in (4) - finger - move. Most J:: may or may not require the second finger and the third no longer be taken. If the factory is in the month, lift the second finger to _. If you move the first finger = finger, the operation can return to the step according to the first finger line = continue to the step heart in this step. If the first hand is no longer detected 128630.doc -58- 200847001

指’拖曳操作可結束DMeans 'drag operation can end D

圖53顯不根據本發明之一項具體實施例的說明性示意動 作操作880。操作88〇可始於步驟882,在此步驟決定是否 债測第_手指。根據系統需要,可識別或不識別第一手 私。例如,某些情形中,若識別指定手指則僅可實施操 作’ *而其他情形中,㈣測任何單—手指則可實施該操 作右偵測一手指,操作可繼續至步驟884,在此步驟實 :·曰向模式。之後,在步驟886中,可決定是否移動手 右移動手扣,步驟888中,可忽略隨後修改符事件。 例如,可忽略額外偵測之手指,僅第一手指之移動可相 :。,後’在步驟_中,可根據移動手指實行指向操 t。:不再偵測到第-手指,指向操作可結束。若第一手 ^ 口疋才呆作可繼續至步驟892,在此步驟尋找修改符事 +。例^,修改符事件可為手指之添加、或藉由新手指作 之不意動作等。之後,步驟m巾,可 定修改蒋塞处 # 疋ώ 1貝列符 。右否,操作可返回步驟886。若是,摔作 :驟896,在此步驟料、統從指向模式切換至拖 果工。右移動第一手指,操作 步驟根千作』、感、,只至步驟898,在此 广據弟—手指移動實行拖良操作。若不㈣測到第一 心’拢曳操作可結束。 圖54顯示根據本發明之一 ♦ 統95 、, I體貝轭例的說明性電子系 李统 $壬何先珂具體實施例組態電子系統。電子 可對應於電腦系統 桌上型電-L w八包細糸統,其包括 知、膝上型電腦、平板電腦或手持式電腦。電子 128630.doc •59· 200847001 系統亦可對應於其他消費性電子冑4,例如手#式裝置, 包括蜂巢式電話、pDA、專用媒體播放器等等。當然,可 將電子系統應用於其他電子裝置。 A圖54中所示之示範性系統950可包括處理器956,其經組 態用以執行指令並執行與系統95〇相關聯之操作。例如, 使用(例如)從記憶體擷取之指令,處理器956可在計算系統 950之組件間控制輸入及輸出資料之接收及操控。可將處 理器956實施於單一晶片、多重晶片或多重電性組件上。 例如,各種架構可詩處理n95G,包括專用錢入式處 理裔、單一用途處理器、控制器、ASIC等等。 大多數情形中,處理器956以及作業系統可運作以執行 電腦代碼並產生及使用資料。作業系統一般係熟知的,不 用詳細說明。例如,作業系統可對應於〇s/2、D〇s、 Unix、Linux、Palm OS等等。作業系統亦可係專用作業系 統,例如可用於限制用途器具型計算裝置。作業系統、其 他電腦代碼及資料可駐留於記憶體區塊958内,其係可操 作地耦合至處理态956。記憶體區塊958可提供位置以儲存 用於系統950之電腦代碼及資料。例如,記憶體區塊%8可 包括唯殯兄憶體(ROM)、隨機存取記憶體(RAM)、硬碟機 及/或類似物。資訊亦可駐留於可抽取儲存媒體上,當需 要時載入或安裝於系統950上。可抽取儲存媒體包括,例 如,CD-ROM、PC卡、記憶卡、軟碟、磁帶、及網路組 件。 系統950亦可包括可操作地耦合至處理器956之顯示裝置 128630.doc -60 - 200847001 968。顯示裝置968可係液晶顯示器(LCD)(例如主動矩陣、 被動矩陣等等)。或者,顯示裝置968可係監視器,例如單 色顯示器、彩色圖形適配器(CGA)顯示器、增強圖形適2 ^(EGA)顯示器、可變圖形陣列(VGA)顯示器、超Vga顯 示器、陰極射'線管(CRT)、#等顯示裝置亦可對應於電浆 顯示器或採用電子墨水實施之顯示器。Figure 53 shows an illustrative schematic action 880 in accordance with an embodiment of the present invention. Operation 88 can begin at step 882 where a decision is made as to whether or not to measure the _ finger. The first hand can be identified or not identified depending on the needs of the system. For example, in some cases, if the specified finger is identified, only the operation '* can be implemented. In other cases, (4) any single-finger can be implemented to perform a right detection of a finger, and the operation can continue to step 884, in this step. Real: · 曰 mode. Thereafter, in step 886, it may be determined whether to move the hand to move the hand button, and in step 888, the subsequent modifier event may be ignored. For example, the extra detected finger can be ignored, and only the movement of the first finger can be: In the step _, the pointing operation can be performed according to the moving finger. : The first finger is no longer detected, and the pointing operation can end. If the first hand is still available, proceed to step 892, where you can find the modifier +. For example, the modifier event can be the addition of a finger, or an unintentional action by a new finger. After that, step m towel, you can modify the Chiang Kai-shek # 疋ώ 1 belle. Right, the operation returns to step 886. If yes, fall: 896, in this step, the system switches from pointing mode to dragging. Move the first finger to the right, and the operation steps are thousands of words, feelings, and only to step 898, where the wide-handed-finger movement is performed. If not (4) the first heart is detected, the drag operation can be ended. Figure 54 shows an illustrative electronic system in accordance with one embodiment of the present invention, a body yoke, and a configuration of an electronic system. Electronics can correspond to computer systems. The desktop-Lw eight-package system includes a laptop, a tablet or a handheld computer. Electronics 128630.doc •59· 200847001 The system can also correspond to other consumer electronic devices 4, such as hand-type devices, including cellular phones, pDA, dedicated media players, and so on. Of course, electronic systems can be applied to other electronic devices. The exemplary system 950 shown in Figure 54 can include a processor 956 that is configured to execute instructions and perform operations associated with the system 95A. For example, processor 956 can control the receipt and manipulation of input and output data between components of computing system 950 using, for example, instructions retrieved from memory. Processor 956 can be implemented on a single wafer, multiple wafers, or multiple electrical components. For example, various architectures can handle n95G, including dedicated money-based processors, single-purpose processors, controllers, ASICs, and more. In most cases, processor 956 and operating system are operable to execute computer code and to generate and use data. Operating systems are generally well known and will not be described in detail. For example, the operating system may correspond to 〇s/2, D〇s, Unix, Linux, Palm OS, and the like. The operating system can also be a dedicated operating system, such as a utility type computing device that can be used for limited use. The operating system, other computer code and data may reside in memory block 958, which is operatively coupled to processing state 956. Memory block 958 can provide a location to store computer code and data for system 950. For example, the memory block %8 may include a syllabary (ROM), a random access memory (RAM), a hard disk drive, and/or the like. The information may also reside on removable storage media and be loaded or installed on system 950 as needed. The removable storage medium includes, for example, a CD-ROM, a PC card, a memory card, a floppy disk, a tape, and a network component. System 950 can also include display device 128630.doc-60 - 200847001 968 operatively coupled to processor 956. Display device 968 can be a liquid crystal display (LCD) (eg, an active matrix, a passive matrix, etc.). Alternatively, display device 968 can be a monitor, such as a monochrome display, a color graphics adapter (CGA) display, an enhanced graphics 2 (EGA) display, a variable graphics array (VGA) display, a super Vga display, a cathode shot 'line Display devices such as tubes (CRT), #, etc. may also correspond to a plasma display or a display implemented with electronic ink.

顯示裝置968可經組態用以顯示圖形使用者介面 (GUI)969,#在系統使用者與作業***或運行於其上之應 用程式間提供易於使用之介面。—般而言,⑽969代^ 具有圖形影像之程式、檔案及操作選項。圖形影像可包括 視窗、攔位、對話框、選單、圖式、独、游標、捲動列 等。可按預定義佈局排列此類影像,或者動態地建立以服 務於使用者所採用之特定動作。操作期間,使用者可選擇 及啟動各種圖形影|,以便起始與其相關聯t功能及任 務:例如,使用者可選擇按鈕,其開啟、_、最小化、 或最大化視窗’或啟動特定程式之圖式。⑽969可在顯 不裔件968上為使用者額外或替代地顯示資訊,例如互動 文子及圖形。 系統950亦可包括可操作地麵合至處理器956之一或多個 輸入衣置97G °輸人裝置97G可經組態用以將資料從外部世 界傳輸至系統950内。輸入裝置97〇可與系統95〇整合,或 者其可係分離組件(或周邊裝置)。某些情形中,可透過有 線連接(例如電纜/埠)將輸入裝置97〇連接至系統95〇。其他 情形中,T透過無線連接將輸人裝置97()連帛至系統95〇。 128630.doc -61 - 200847001 貧料鏈路可對應於PS/2、USB、IR、RF、藍芽等 例如’輸人|置97G可用於實行追蹤,以及在顯示器968 3 ; GUI 969作出選擇。輸入裝置97〇亦可用於在系統 95〇中發出命令。 y項具體實施例中,輸入裝置970可包括觸控感測裝 置’其經組態用以從使用者之觸控接收輸人並將此資訊傳Display device 968 can be configured to display a graphical user interface (GUI) 969, which provides an easy to use interface between the system user and the operating system or applications running thereon. In general, (10) 969 generation ^ has graphics, programs, files and operating options. Graphical images can include windows, docks, dialogs, menus, schemas, uniques, cursors, scrolling columns, and more. Such images may be arranged in a predefined layout or dynamically created to serve a particular action taken by the user. During operation, the user can select and launch various graphics images to initiate t functions and tasks associated with them: for example, the user can select a button that opens, _, minimizes, or maximizes the window' or launches a particular program. The pattern. (10) 969 may display additional or alternative information, such as interactive text and graphics, for the user on the display device 968. System 950 can also include one or more of the operatively coupled to processor 956. The input device 97G can be configured to transfer data from the outside world to system 950. The input device 97A can be integrated with the system 95, or it can be a separate component (or peripheral device). In some cases, the input device 97A can be connected to the system 95A via a wired connection (e.g., cable/twist). In other cases, T connects the input device 97() to the system 95 via a wireless connection. 128630.doc -61 - 200847001 The poor link can correspond to PS/2, USB, IR, RF, Bluetooth, etc. For example, 'Input' 97G can be used to implement tracking, and on display 968 3; GUI 969 to make a selection. Input device 97A can also be used to issue commands in system 95. In a specific embodiment, the input device 970 can include a touch sensing device configured to receive input from a user's touch and transmit the information.

廷至處理态956。例如,觸控感測裝置可對應於觸控板、 觸控螢幕及/或觸敏外罩。 夕丨月开&gt;/中觸控感測裝置可辨識觸控(或接近觸控), 、觸控面板表面上之觸控的位置及大小。觸控感測構件 可向處理器956報告觸控,處理器州可根據其程式化解譯 觸控二例如’處理器956可根據特定觸控起始任務。專用 處理益可用於在本機處理觸控,並減小對於電腦系統主要 處理器之需求。觸控感測裝置可基於感測技術,包括但不Tingzhi processing state 956. For example, the touch sensing device may correspond to a touch panel, a touch screen, and/or a touch sensitive cover. The touch sensing device can recognize the touch (or proximity touch), the position and size of the touch on the surface of the touch panel. The touch sensing component can report the touch to the processor 956, and the processor state can interpret the touch according to its programming. For example, the processor 956 can initiate a task according to a particular touch. Dedicated processing benefits can be used to handle touch on the unit and reduce the need for the main processor of the computer system. The touch sensing device can be based on sensing technology, including but not

例如 等。 限於電4感測、電阻感測、表面聲波感測、壓力感測、光 學感測及/或類似物。 另外’觸控感測構件可基於單點感測或多點感測。單點 感測僅能夠區別單-觸控,❿多點感測能夠區別同時發生 的多個觸控。或者或此外’可使用近接感測裝置。近接感 測褒置能夠感測空間内而非觸控表面上的物件。 -項具體實施例中,輸入裝置97〇可包括多點輪入裝 置。多點輸人裝置具有優於傳統單點裝置之優點,即呈可 同時區別-個以上個物件(手指)。例如,本文可使用:多 12863〇.d〇c * 62 - 200847001 點裝置在美國專利申請案第10/840,862及1 1/015,434號詳細 顯示及說明,其均以提及方式併入本文。 根據本發明之一項具體實施例,系統950可係設計成辨 識應用於一或多個輪入裝置970(特別係觸控裝置)之示意動 作985 ’並根據示意動作985控制系統95〇之方面。如先前 所述’可將示意動作定義為與輸入裝置之格式化互動,其 係映射至一或多個特定計算操作。示意動作985可透過不 同手完成,更特定言之係手指動作。或者或此外,示意動 作可採用觸控筆完成。所有該等情形中,輸入裝置97〇之 至少一個,例如觸控裝置,可接收示意動作985,並且處 理為956可執行指令以執行與示意動作985相關聯之操作。 此外$己炫體區塊958可包括示意動作操作程式988,其可 能係作業系統之部分或分離應用程式。示意動作操作程式 988可包括組指令,其辨識示意動作985之出現並通知示 意動作985之一或多個軟體代理及/或回應示意動作985而 採取何種動作。 當使用者實行一或多個示意動作時,輸入裝置970可轉 送不思動作貧訊至處理器956。使用來自記憶體%8之指 令,更特定言之係示意動作操作程式988,處理器9%可解 澤不意動作985並根據示意動作985控制系統95〇之不同組 件,例如記憶體958、顯示器968。示意動作可被識別 為命令,其用於在儲存於記憶體958内之應用程式中實行 動作、修《:顯示於顯968上之GUI物件、修改储存於記 憶體958内之資料、及/或在1/〇裝置内實行動作。例如,命 I28630.doc •63- 200847001 令可與縮放、平移、捲動 關聯。作為其他範例,命人轉、調整大小等等相 案或文件、檢視選單、作:選:、:動特定程式、開啟檔 統、允許授權個體存取電、统執订指令、登入電腦系 者之較佳電腦桌面配署,… 制區域、載入與使用 物相關聯。。置相關聯的使用者設定播及/或類似 一項具體實施例中,接觸輸入裝置 入模式。例如D單 4目可心不輸 行追蹤之需要,如it4 扣作出之皁一觸控可指示實 群組作出之多重觸二?游:移動或選擇’而(例如)手指 於實施示意動作之手心:;:其他示意動作之需要。用 可指示第-輸入模二個;;二化:例如,兩個手指 或者,任何數目之手,二可“第三輸入模式等。 式,装可勺紅 才曰即一個以上’可用於相同輸入模 用…:多個示意動作控制。手指方位可相同地 用於表不期望模式。 否(例如廓以根據使用者是否是 ^㊉^允許不同模型操作。 m具體實施例中’單—手指可起始 兩個或兩個以上彼此緊密靠近之手指可起始捲動或 使用者可藉由二:::手指間容易的切換,即, 模式間切換。此1有:Γ 手指非常容易地在 追 /、有比其他模式切換形式更直觀的優點。 動时1曰’匕可藉由使用者在觸控感測裝置之觸敏表面上移 解:车?曰控制游標移動。觸控感測裝置之感測器配置可 動作並產生用於在顯示器上產生游標之對應移動 128630.doc -64- 200847001 的信號。捲動期門 _ ^ θ ’可猎由使用者在觸控感測裝置之觸敏 二亡Τ動兩個手指控制螢幕移動。當沿垂直方向移動組 手私恰可將動作解譯為垂直捲動事件,當沿水平方向 、 口手私4,可將動作解譯為水平捲動事件。相同情 况了用於平移’儘管平移可發生於除水平及垂直方向外的 所有方向上。For example, etc. Limited to electrical 4 sensing, resistance sensing, surface acoustic wave sensing, pressure sensing, optical sensing, and/or the like. Additionally, the touch sensing member can be based on single point sensing or multi-point sensing. Single-point sensing can only distinguish between single-touch, and multi-point sensing can distinguish multiple touches that occur simultaneously. Alternatively or additionally, a proximity sensing device can be used. The proximity sensing device is capable of sensing objects in space rather than on the touch surface. In a particular embodiment, the input device 97A can include a multi-point wheeling device. The multi-point input device has the advantage over the conventional single-point device, that is, it can simultaneously distinguish more than one object (finger). For example, the following is a detailed description and illustration of U.S. Patent Application Serial Nos. 10/840,862 and 1 1/015,434, each of which is incorporated herein by reference. In accordance with an embodiment of the present invention, system 950 can be designed to recognize schematic actions 985' applied to one or more wheeling devices 970 (particularly touch devices) and control system 95 in accordance with schematic action 985. . As indicated previously, a schematic action can be defined as a formatted interaction with an input device that maps to one or more particular computing operations. The gesture 985 can be done by different hands, more specifically finger movements. Alternatively or additionally, the gesture can be accomplished using a stylus. In all such cases, at least one of the input devices 97, such as a touch device, can receive a gesture 985 and process 956 executable instructions to perform the operations associated with the gesture 985. In addition, the $ illusion block 958 can include a gesture action program 988 that can be part of the operating system or separate application. The gesture action program 988 can include group instructions that identify the occurrence of the gesture 985 and notify the one or more software agents of the action 985 and/or the action 985 to take action. When the user performs one or more gestures, the input device 970 can forward the message to the processor 956. Using instructions from memory %8, more specifically, the action program 988, the processor 9% can delineate the action 985 and control the various components of the system 95 according to the gesture 985, such as the memory 958, display 968 . The gesture may be identified as a command for performing an action in an application stored in the memory 958, modifying the GUI object displayed on the display 968, modifying the data stored in the memory 958, and/or The action is performed in the 1/〇 device. For example, the I28630.doc •63- 200847001 command can be associated with zoom, pan, scroll. As another example, the person to turn, resize, etc., or the menu, view the menu, make: select:,: move the specific program, open the file system, allow authorized individuals to access the electricity, order the order, log in to the computer system The preferred computer desktop distribution department, ... the area, loading and use of the object. . In an associated user setting play and/or similar embodiment, the touch input device enters the mode. For example, D single 4 eyes can not be traced to the needs of tracking, such as the soap made by it4 button can indicate the multiple touches made by the real group? Travel: move or select 'and (for example) the finger to implement the gesture Palm::: The need for other gestures. It can be used to indicate the first-input mode; two; for example, two fingers or any number of hands, two can be "third input mode, etc., can be used to match the red one or more" can be used for the same The input mode is controlled by: multiple gestures. The finger orientation can be used identically to indicate the undesired mode. No (for example, the profile is allowed to operate according to whether the user is a ^^^^. A finger that can start two or more close to each other can initiate scrolling or the user can easily switch between two fingers by using two::: between modes. This 1 has: 手指 Fingers are very easy In chasing /, there is a more intuitive advantage than other mode switching modes. The movement can be moved by the user on the touch-sensitive surface of the touch sensing device: the car 曰 controls the movement of the cursor. The sensor configuration of the measuring device is operable to generate a signal for the corresponding movement 128630.doc-64-200847001 for generating a cursor on the display. The scrolling gate _ ^ θ ' can be hunted by the user in the touch sensing device Touch sensitive two dead and shake two finger control Curtain movement. When moving the group in the vertical direction, the action can be interpreted as a vertical scrolling event. When the horizontal direction and the mouth are private, the action can be interpreted as a horizontal scrolling event. The same situation is used for translation. 'Although translation can occur in all directions except horizontal and vertical.

ΓΓΓ述之具體實施例可藉由硬體、軟體或硬體及軟體 似、且:Λ &amp;。亦可將軟體具體化為電腦可讀取媒體上的電 可項取代碼。電腦可讀取媒體係任何資料儲存裝置,其 :儲存之後可由電腦系統讀取之資料。電腦可讀取媒體2 乾例包括唯讀記憶體、隨機存取記憶體、CD_R〇M、 磁帶、光學資料儲存裝置、及載波。亦可將電腦可 嗔取媒體分佈於網路耦合電腦系統上,以便儲存及以分佈 式方式執行電腦可讀取代碼。 刀 雖然已就數個較佳具體實施例說明本發明,本發明之$ 圍内存在修改、置換及等效物。例如,儘管本發明主要係 针對觸控裝置,應注意本發明亦可採料接感測裝置來# 施。此外,可實施方法以自動同步橫跨多個裝立只 .* t 不思動 作偏好。例如,當將掌上裝置連接 立 设王王狨電腦時。亦應注 思,本發明之方法及裝置存在許多替代實施方式。伤〜 儘管本發明主要係針對與多點感測裝置相關聯1方=如, 注意,某些方法可係針對單點感測裝置。 / ,應 口此以下隨附φ 請專利範圍係解釋為包括本發明之真實精1〒 有此類更改、置換、及等效物。、以的所 128630.doc -65 - 200847001 圖式簡單說明】 、、、口 〇 附圖參考上述詳細說明,將容易地理解本發明,其 中相似參考數字指定相似結構元件,以及其中: 圖1顯示根據本發明之一項具體實施例的說明性示竟 作控制操作。 圖2顯示根據本發 明 作 之一項具體實施例的說明性控制操 作 作 作 圖3 s苞- 4 μ不根據本發明之一項具體實施例的說明性控制操 0 圖4蕊苜—4曰 ”、、/、根據本發明之一項具體實施例的說明性控制操 〇 圖5辱苜—丁 · 4曰4·占 .、、、不根據本發明之一項具體實施例的說明性控制操 圖6顯示根據本發 作 明之一項具體實施例的說明性控制操 圖7顯 作操作 示根據本發明之一項具體實施·例的說明性示意動 圖8顯示插滅·士 作操作 據本發明之一項具體實施例的說明性示意動 圖9 _ + 4日 ”、、’、Χ據本發明之一項具體實施例的說明性示意動 作操作。 圖1 〇顯示根摅太八 作操作 據本每明之一項具體實施例的說明性示意動 圖11顯示根擔士 n 據本^明之一項具體實施例的說明性示意動 128630.doc -66 - 200847001 作操作。 圖’員示根據本發明之一項具體實施例的說明性示意動 作操作。 圖13顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖14顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖1 5顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖16顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖17顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖1 8顯不根據本發明之一項具體實施例的說明性示意動 作操作。 圖19 ·、、、員不根據本發明之一項具體實施例的說明性示意動 作操作。 gj ^ Q晷茛 _ 回Μ不根據本發明之一項具體實施例的說明性控制操 作。 圖1 .、、、員不根據本發明之一項具體實施例的說明性控制操Specific embodiments of the description may be by hardware, software or hardware and software, and: &amp; The software can also be embodied as an electrical item code on a computer readable medium. The computer readable media is any data storage device that stores data that can be read by the computer system after storage. Computer readable media 2 Examples include read-only memory, random access memory, CD_R〇M, magnetic tape, optical data storage, and carrier. Computer-capable media can also be distributed over network-coupled computer systems to store and distribute computer-readable code in a distributed manner. Knife Although the invention has been described in terms of several preferred embodiments, modifications, substitutions and equivalents are present in the invention. For example, although the present invention is primarily directed to a touch device, it should be noted that the present invention can also be used with a sensing device. In addition, methods can be implemented to automatically synchronize across multiple installations only.* t not thinking about preferences. For example, when connecting a handheld device to a king computer. It should also be noted that there are many alternative embodiments of the method and apparatus of the present invention. Injury ~ Although the present invention is primarily directed to a party associated with a multi-point sensing device = as noted, some methods may be directed to a single point sensing device. / , should be attached to the following φ. The scope of the patent is to be construed as including the true and precise features of the present invention. BRIEF DESCRIPTION OF THE DRAWINGS The present invention will be readily understood by reference to the above detailed description, wherein like reference numerals designate similar structural elements, and wherein: Figure 1 shows An illustrative operation in accordance with an embodiment of the present invention is contemplated as a control operation. 2 shows an illustrative control operation in accordance with an embodiment of the present invention. FIG. 3 s - 4 μ is an illustrative control operation not according to an embodiment of the present invention. FIG. 4 </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> <RTIgt; 6 shows an illustrative control operation according to a specific embodiment of the present invention. FIG. 7 shows an operation diagram according to an embodiment of the present invention. FIG. 8 shows an exemplary operation. Illustrative schematic diagram of a particular embodiment of the invention is an illustrative illustrative operational operation of a particular embodiment of the present invention. 1 shows an illustrative schematic diagram of a specific embodiment of the present invention. FIG. 11 shows an illustrative diagram of a specific embodiment of the present invention. 128630.doc-66 - 200847001 Operation. Figure </ RTI> shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 13 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 14 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 15 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 16 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 17 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 18 shows an illustrative schematic operational operation in accordance with an embodiment of the present invention. Figure 19 is an illustrative schematic operational operation of an embodiment of the present invention. Gj ^ Q晷茛 _ Μ An illustrative control operation that is not in accordance with a particular embodiment of the present invention. Figure 1, . . , an illustrative control operation not according to a specific embodiment of the present invention

作。 A 作圖2,、、、員不根據本發明之一項具體實施例的說明性控制操 圖23顯不根據本發明之一項具體實施例的說明性控制操 128630.doc -67- 200847001 作。 圖24顯示根據本發明之-項具體實施例的說明性控制操 作。 蹄 圖25顯不根據本發明之一項具體實施例的說明性控制 作。 工罘 圖26顯不根據本發明之一項具體實施例的說明性控制 作。 ’、 圖27顯示根據本發明之一項具體實施例的說明性控制操 作。 ’、 圖28顯不根據本發明之一項具體實施例的說明性控制操 圖2 9 A至G顯示根據本發明之一項具體實施例的說明性 控制面板之圖式。 圖30顯不根據本發明之一項具體實施例的說明性控制操 作。 ’、 圖3 1顯不根據本發明之一項具體實施例的說明性示意動 作映射方法。 ~ 圖32顯不根據本發明之一項具體實施例的說明性 作映射方法。 』 圖33顯不根據本發明之一項具體實施例的說明性示咅動 作回饋方法。 ^ 、圖34顯不根據本發明之一項具體實施例的示意動作檢視 視窗之說明性圖式。 圖35顯不根據本發明之一項具體實施例經由觸控螢幕實 128630.doc •68· 200847001 施的說明性鍵盤方法。 施例的說明性示意動 圖3 6顯不根據本發明之一項具體實 作設計程序。 、 明之一項具體實施例用於分級弦之說 圖37顯示根據本發 明性矩陣。 圖38顯示根據本發 之項具體實施例用於針對一此择 作分級操作頻率的說明性矩陣。 二操Work. A, FIG. 2, and an illustrative control operation diagram not according to a specific embodiment of the present invention is shown in accordance with an illustrative control operation of a specific embodiment of the present invention 128630.doc-67-200847001 . Figure 24 shows an illustrative control operation in accordance with an embodiment of the present invention. Hoof Figure 25 shows an illustrative control in accordance with an embodiment of the present invention. Work Figure 26 shows an illustrative control in accordance with an embodiment of the present invention. Figure 27 shows an illustrative control operation in accordance with an embodiment of the present invention. Figure 28 shows an illustrative control panel in accordance with an embodiment of the present invention. Figures 9A through G show diagrams of illustrative control panels in accordance with an embodiment of the present invention. Figure 30 shows an illustrative control operation in accordance with an embodiment of the present invention. Figure 31 shows an illustrative schematic action mapping method in accordance with an embodiment of the present invention. ~ Figure 32 shows an illustrative mapping method in accordance with an embodiment of the present invention. Figure 33 shows an illustrative dynamic feedback method in accordance with an embodiment of the present invention. Figure 34 shows an illustrative diagram of a schematic motion view window in accordance with an embodiment of the present invention. Figure 35 illustrates an illustrative keyboard method implemented via a touch screen 128630.doc • 68· 200847001, in accordance with an embodiment of the present invention. An illustrative schematic of the embodiment of the invention is shown in accordance with a particular embodiment of the present invention. A specific embodiment of the invention for grading strings Fig. 37 shows a matrix according to the present invention. Figure 38 shows an illustrative matrix for classifying operating frequencies for one such selection in accordance with an embodiment of the present invention. Second operation

圖39顯示根據本發 作。 項具體實施例的說明性控制操 圖40顯示根據本發 作。 項具體實施例的說明性控制操 圖41顯示根據本發 — 作 項具體實施例的說明性控制操 圖42顯示根據本發明之一 之說明性圖式 項具體實施例的示意動作映射Figure 39 shows the release according to the present invention. An illustrative control scheme of an embodiment is shown in accordance with the present invention. Illustrative Control Operation of a Specific Embodiment FIG. 41 shows an illustrative control map in accordance with an embodiment of the present invention, showing a schematic motion map of an illustrative embodiment of an exemplary embodiment of the present invention.

之圖式 圖44顯示根據本發明 之圖式。 圖45顯示根據本發明之一 圖式。 之項具體實施例的示意動作映射 具體實施例的示意動作竿 之 圖頁丁根據本發明之_項具體實施例的示意動作集 圖式 圖4 7顯示根據本發 之 明之一項具體實施例的示意動作集 之 128630.doc -69- 200847001 圖式。 之一項具體實施例的示意動作集之 圖48顯示根據本發明 圖式。Figure 44 shows a diagram in accordance with the present invention. Figure 45 shows a diagram in accordance with the present invention. Illustrated action map of a specific embodiment of the present invention, a schematic action diagram of a specific embodiment of the present invention, and a schematic action set diagram of a specific embodiment according to the present invention, FIG. 47 shows a specific embodiment according to the present invention. Schematic action set 128630.doc -69- 200847001 schema. A schematic set of actions of a particular embodiment of Figure 48 shows a diagram in accordance with the present invention.

圖49顯示根據本發明之— 作操作。 圖50顯示根據本發明之一 作操作。 圖5 1顯示根據本發明之一 作操作。 圖5 2顯示根據本發明之一 作操作。 圖5 3顯示根據本發明之一 作操作。 圖5 4顯示根據本發明之一 明性方塊圖。 項具體實施例的說明性示音動 項具體實施例的說明性示意動 項具體實施例的說明性示音、動 項具體實施例的說明性示意動 項具體實施例的說明性示意動 項具體實施例的電子系統之說Figure 49 shows the operation in accordance with the present invention. Figure 50 shows an operation in accordance with one of the present invention. Figure 51 shows the operation in accordance with one of the present invention. Figure 5 2 shows operation in accordance with one of the present invention. Figure 53 shows the operation in accordance with one of the present invention. Figure 5 4 shows a block diagram of one of the aspects of the present invention. Illustrative vocabulary of an embodiment of the present invention. Illustrative vocabulary of an embodiment of the present invention. The electronic system of the embodiment

【主要元件符號說明】 550 控制面板 552 示意動作映射控制 554 輸入功能選單 556 示意動作集區段 558 弦選單 660 示意動作檢視視窗 662 分離方框 664 示意動作事件 128630.doc -70· 200847001 667 命令 950 電子系統 956 處理器 958 記憶體區塊 968 顯示裝置 969 圖形使用者介面 970 輸入裝置 985 示意動作 988 示意動作操作程式 128630.doc -71 ·[Main component symbol description] 550 Control panel 552 Schematic action mapping control 554 Input function menu 556 Schematic action set section 558 Chord menu 660 Schematic action view window 662 Separation block 664 Schematic action event 128630.doc -70· 200847001 667 Command 950 Electronic system 956 processor 958 memory block 968 display device 969 graphical user interface 970 input device 985 gesture action 988 schematic action program 128630.doc -71

Claims (1)

200847001 、申請專利範圍: 一種電子系統,其包含: 一多點感測裝置,其提供用於從一或多個物件接收輸 入之一多點感測區域; 一示意動作模組,其經組態用以決定用於藉由該多點 感=裝置之該多點感測區域接收之—給定輸人配置的一 不思動作集,用以監視用於包括於該示意動作集内之一 或多個示意動作集的該給定輸人配置,以及用以在採用 該輸入配置實行該示意動作事件時起始與—示意動作事 件相關聯之輸入動作。 其中該多點感測裝置係一觸控 2·如請求項1之電子系統 感測裝置。 其中該多點感測裝置係一近接 3·如請求項1之電子系統 感測裝置。 如明求項1之電子系統,苴中 中糟由硬體、軟體或硬體及 體之—組合實施該示意動作模組。 5. =5項1之電子线,其中該電子系、統與—電腦系統 手二其包括桌上型電腦、膝上型電腦、平板電腦或 予持式電腦。 6. 如請求項1之電子夺统,並士 性雷早# 丨中該電子系統與手持式消費 衣置相關聯,該等手持式消費性電子 7 :式電話、舰、或專用媒體播放器。 7. 如請求们之電子系統 : 個因 、中該不意動作集亦由一或多 …該一或多個因素包括該輸入配置在該感測 128630.doc 200847001 區域内之位置、一開啟應 -m ^ 狂式、该開啟應用程式之一 狀::輸入功能性、或額外輪入特徵。 '之 /、中該等輸入動作係狀態命 8·如請求項1之電子系統 令0 八中該等輸入動作係操控命 其中藉由該輸入配置實行該等 其中藉由該輸入配置之一子集 9·如請求項1之電子系統 令0 10·如請求項1之電子系統 示意動作事件。 11 ·如請求項1之電子系統 實行該等示意動作事件 12.如請求項i之電子系統,其中決 疋Θ不思動作集包括辨 鐵一特疋輸入配置,以及其中龄 皿視該給疋輸入配置包括 辨識與該示意動作集相關聯之示意動作事件。 13·如請求項12之電子系統’其中辨識該等示意動作事件包 括·分析該輸人配置之各輸人的特徵;識別與該輸入配 置相關聯之一特定圖案;以及參考立 &gt;芩鑌不思動作集以查看 該特定圖案是否匹配與該示意動作集相關聯之一示音 作事件。 ’ W 14 ·如請求項1之電子系統,其中示咅 丁丁思動作事件與觸控相關 聯,其可係選自運動、敲擊、壓力、及停留。 b.如請求項14之電子系統,其中運動係選自滑動、擦過、 輕敲、及旋轉。 16·如晴求項14之電子糸統’其中運動传雨摘七 咬籾你兩個或兩個以上輸 入之間的收縮或膨服。 128630.doc 200847001 U·如請求項14之電子系統,其中敲擊係選自1/2敲擊、完全 敲擊、多次敲擊、快速敲擊及缓慢敲擊。 1 8 ·如請求項14之電子系統,其中壓力與重及輕按壓相關 聯。 19·如請求項14之電子系統,其中停留與一長或短固定暫停 相關聯。 20.如請求項14之電子系統’其中該示意動作集由不同示意 動作事件及不同命令之一矩陣組成。 21·如請求項20之電子系統,其中該示意動作集由9個不同 示意動作事件及9個不同命令之一 3 x 3矩陣組成。 22·如凊求項2〇之電子系統,其中該9個不同示意動作事件 包括旋轉CCW、旋轉CW、擴展、收縮、向左擦過、向 右擦過、向上擦過、向下擦過、及敲擊。 23. —種示意動作控制方法,其包含: 门日守感測一感測區域内之多個點; 當在該感測區域内债測一或多個·點時決定一弦,該弦 係該感測區域内之一特定點配置; 、、定不^動作集,其將命令與一或多個示意動作事 件相關聯; 監視用於#意動作事件之點;以及 右辨識一示 聯之命令。 “動作事件,則實行與示意動作事件相 關 其中該等點與發生於一觸敏表面上 2 4 ·如请求項2 3之方法 之觸控相關聯。 128630.doc 200847001 其中該等點與發生於-表面特徵上 其中採用-或多個手之全部或部分 其t該弦係基於若干未知點。 其中該弦係基於已知點之一特定配 25·如請求項23之方法 之接近觸控相關聯 26·如請求項23之方法 完成該弦。 27.如請求項23之方法 28·如請求項23之方法 置。 2求項23之H其中該弦係基㈣等點是否係彼此 罪近、位於一中性位置或分散。 3〇·如明求項23之方法,其中該等弦係基於該等接觸是否係 彼此相鄰或偏移。 其中該等弦係基於其是否來自左或 31.如請求項23之方法 右手。 32·如請求項23之方法,i φ冰宁分护a t 八中决疋該弦包括分析藉由觸控或 接近觸控該感測區域之物件建 初忏遝立的影像,以及從該影像 辨識一特定點配置。 33·如請求項23之方法,复申呤+音龢 八T該不思動作集係基於該弦。 3 4·如請求項33之方法,i中兮千音毹 ,、中違不思動作集亦係基於額外因 素’其包括該弦之位詈、閲敌雍 1開啟應用転式、該應用程式之 狀愍或更多、及/或該等點之特徵。 35·如請求項23之方法,i中哕音 T 4不思動作集與一輸入功能性 相關聯。 36·如請求項35之方法, 具中邊輸入功能性係選自導覽操 作、檔案操作、編輯捶你、蛤满 、 W梯作、檢視彳呆作、格式化操作、工 128630.doc 200847001 具操作、或網頁瀏覽操作。 37:=項36之方法’其中該等命令係選自狀態命令及操 38.^未項23之方法,其中辨識示意動作事件包括:分析 特^,Μ—特定圖案;及參考該示意動作集 意動作事件清單,若-圖案匹配該示意動作集之 不思動作事件,則假定該圖案為該示意動作事件。 39· ^請求項38之方法,其中該等點之該等特徵包括:第一 等級考里’其包括運動、敲擊、壓力或停留;以及第二 等級考ΐ,包括速度、方向、方位、尺寸、持續時間、 形狀、或長度。 復如請求項23之方法,其中示意動作事件與運動、敲擊、 壓力、及停留相關聯。 41.如請求項40之方法,其中運動係選自滑動、擦過、輕 敲、及旋轉。 42·如請求項40之方法,其中運動係兩個或兩個以上輸入之 間的收縮或膨脹。 43·如請求項40之方法,其中敲擊係選自%敲擊、完全敲 擊、多次敲擊、快速敲擊及緩慢敲擊。 44·如請求項40之方法,其中壓力與重及輕按壓相關聯。 45.如請求項40之方法,其中停留與一長或短固定暫停相關 聯。 46·如請求項23之方法,其中該示意動作集由不同示意動作 事件及不同命令之一矩陣組成。 I28630.doc 200847001 47.如凊求項46之方法,其中該示意動作集由9個不同示意 動作事件及9個不同命令之一3 χ 3矩陣組成。 Μ.如請求項47之方法,其中該9個不同示意動作事件包括 旋轉CCW、旋轉CW、擴展、收縮、向左擦過、向右擦 過、向上擦過、向下擦過、及敲擊。 49.如請求項23之方法,其進一步包括決定是否已實行一切 換事件,一切換事件指擱置該操作或起始一弦變化之一 事件。 5〇·如請求項23之方法,其中藉由以下瘂 τ稭田以下刼作之一實施該切換 事件:a)在-預定時間量内從感測區域移除所有點;b) 改變該弦;〇)從感測區域添加/減去點;d)在一預定時間 量内保持該等點固定;eM貞測一特定按鍵輸人或按紐點 選,或f)彳貞測一示意動作事件。 5 1 · —種控制操作,其包含: 感測一觸控或接近觸控; 立若感測到-觸控,為該觸控決定一示意動作集,該示 意動作集包括用於引發或起始—命令之—或多個示 作事件; 針對與該示意動作集相關聯 丨剛%”之不思動作事件監視該 觸控;以及 當實行與該示意動作集相關, 一九 木相關如之一不意動作事件時 始^命令。 52·如請求項51之操作,其中嗲 甲鑌不思動作集取決於觸控 徵、觸控位置、開啟應用程.、 狂式及/或應用程式模式。 128630.doc 200847001 5 3 ·如明求項5 1之操作,其中經由一觸控裝置,例如觸控 板、觸控螢幕或觸敏外罩,偵測該觸控或接近觸控。 54· —種示意動作操作,其包含·· 監視一觸控運動; 在第一與第二狀態間區分該觸控運動; 若該觸控運動與該第一狀態相關聯,則實行一第一動 作;以及 若該觸控運動與該第二狀態相關聯,則實行一第二動 作0 55. 如請求項54之示意動作操作,其中該觸控運動係一滑動 、及其中該苐一狀悲係一輕敲,並且該第二狀態 係一擦過。 56· —種控制操作,其包含: 一輸入 提供一第一輸入裝置及一第二輸入裝置,該第 裝置包括一觸敏表面;200847001, Patent Application Range: An electronic system comprising: a multi-point sensing device providing a multi-point sensing area for receiving input from one or more objects; a schematic action module configured Determining, for determining, by the multi-point sense=the multi-point sensing region of the device, a set of inactions for a given input configuration for monitoring for inclusion in one of the set of gestures or The given input configuration of the plurality of gesture sets and the input action associated with the initial gesture event when the gesture action is performed using the input configuration. Wherein the multi-point sensing device is a touch control device, such as the electronic system sensing device of claim 1. Wherein the multi-point sensing device is in close proximity to the electronic system sensing device of claim 1. According to the electronic system of claim 1, the schematic operation module is implemented by a combination of hardware, software or hardware and body. 5. = 5 items 1 of the electronic line, where the electronic system, system and computer system, including the desktop computer, laptop, tablet or holding computer. 6. If the electronic item of claim 1 is taken, and the electronic system is associated with the handheld consumer clothing, the handheld consumer electronic 7: type telephone, ship, or dedicated media player . 7. If the electronic system of the requester: the cause, the unintentional action set is also one or more... The one or more factors include the position of the input configuration in the area of the sensing 128630.doc 200847001, an opening should be - m ^ mad, the one of the open applications: input functionality, or extra rounding features. ' / /, the input action state is 8; such as the electronic system of claim 1 is 0, the input action is controlled by the input configuration, wherein one of the input configurations is implemented by the input configuration Set 9. The electronic system of claim 1 is 0. 10. The electronic system of claim 1 indicates an action event. 11. The electronic system of claim 1 implements the illustrated action events. 12. The electronic system of claim i, wherein the set of actions includes a configuration of the iron and the special input, and the middleware is deemed to be The input configuration includes identifying a gesture action event associated with the set of gestures. 13. The electronic system of claim 12, wherein identifying the graphical action events comprises: analyzing characteristics of each of the input configurations; identifying a particular pattern associated with the input configuration; and referencing a &gt; The action set is not thought to see if the particular pattern matches one of the voice events associated with the set of gestures. W 14 The electronic system of claim 1 wherein the Ding Ding action event is associated with touch, which may be selected from the group consisting of exercise, tapping, stress, and stay. b. The electronic system of claim 14, wherein the motion is selected from the group consisting of sliding, wiping, tapping, and rotating. 16·如晴求14的电子糸', where the movement is raining seven, biting the contraction or expansion between two or more of your inputs. 128 430.doc 200847001 U. The electronic system of claim 14, wherein the tapping is selected from the group consisting of a 1/2 tap, a full tap, a multiple tap, a quick tap, and a slow tap. 1 8 The electronic system of claim 14, wherein the pressure is associated with a weight and a light press. 19. The electronic system of claim 14, wherein the stay is associated with a long or short fixed pause. 20. The electronic system of claim 14 wherein the set of schematic actions consists of a matrix of different illustrative action events and different commands. 21. The electronic system of claim 20, wherein the set of gestures consists of nine different gesture events and one of three different commands, a 3 x 3 matrix. 22. The electronic system of claim 2, wherein the nine different gesture events include rotating CCW, rotating CW, expanding, contracting, rubbing to the left, wiping to the right, rubbing upward, wiping down, and tapping. 23. A schematic motion control method, comprising: sensing a plurality of points in a sensing area; determining a string when the one or more points are measured in the sensing area, the string a specific point configuration within the sensing area; , an action set that associates the command with one or more gesture events; a point for monitoring the #意 action event; and a right recognition command . The "action event" is associated with a gesture event in which the points are associated with a touch occurring on a touch-sensitive surface, such as the method of claim 2. 128. - the surface feature in which - or all or part of the plurality of hands are employed, t the string is based on a number of unknown points. wherein the string is based on one of the known points. 25 is as close as touch related to the method of claim 23. 26. The method of claim 23 completes the chord. 27. The method of claim 23 is as set forth in claim 23, and the method of claim 23, wherein the chord base (four) is equal to each other. The method of claim 23, wherein the chord is based on whether the contacts are adjacent or offset from each other, wherein the chords are based on whether they are from the left or 31. The method of claim 23 is the right hand. 32. The method of claim 23, i φ冰宁分护 at 八中疋 The string includes analyzing the object of the sensing area by touch or proximity touch a standing image and a recognition from the image Fixed-point configuration 33. As in the method of claim 23, the 呤+ 音 and the eight-T do not think about the action set based on the chord. 3 4·If the method of claim 33, i 兮 毹 毹,, 中 中The non-thinking action set is also based on additional factors 'which include the position of the string, the application of the enemy 1 open application, the status of the application or more, and/or the characteristics of the points. 35. The method of item 23, wherein the voice T 4 motion control set is associated with an input function. 36. The method of claim 35, wherein the middle input function is selected from a navigation operation, a file operation, and an editing. You, 蛤满, W ladder, view 彳 、, formatting operation, work 128630.doc 200847001 with operations, or web browsing operations. 37: = Item 36 method 'where the commands are selected from the state command and operation 38. The method of claim 23, wherein the identifying the action event comprises: analyzing the feature, the Μ-specific pattern; and referring to the list of the action action set event events, if the pattern matches the action action event of the gesture set, It is assumed that the pattern is the schematic action event. 9. The method of claim 38, wherein the features of the points include: a first level of test, which includes motion, tapping, pressure, or dwell; and a second level of test, including speed, direction, orientation, The method of claim 23, wherein the action event is associated with motion, tapping, stress, and dwell. 41. The method of claim 40, wherein the motion is selected from a sliding 42. The method of claim 40, wherein the motion is contraction or expansion between two or more inputs. 43. The method of claim 40, wherein the tapping is selected from the group consisting of a % tap, a full tap, a multiple tap, a quick tap, and a slow tap. 44. The method of claim 40, wherein the pressure is associated with a weight and a light press. 45. The method of claim 40, wherein the stay is associated with a long or short fixed pause. 46. The method of claim 23, wherein the set of gestures consists of a matrix of different gesture events and a different command. 47. The method of claim 46, wherein the set of gestures consists of nine different schematic action events and one of three different commands, a 3 χ 3 matrix. The method of claim 47, wherein the nine different gesture events include rotating CCW, rotating CW, expanding, contracting, rubbing left, wiping to the right, rubbing up, swiping down, and tapping. 49. The method of claim 23, further comprising deciding whether a change event has been performed, a switch event being an event that suspends the operation or initiates a change in a string. The method of claim 23, wherein the switching event is performed by one of the following operations: a) removing all points from the sensing area for a predetermined amount of time; b) changing the string ;)) add/subtract points from the sensing area; d) keep the points fixed for a predetermined amount of time; eM measures a specific button input or button selection, or f) speculates a gesture event. 5 1 · a control operation, comprising: sensing a touch or proximity touch; directly sensing a touch, determining a set of gestures for the touch, the set of gestures included for triggering or starting Start-command-- or multiple display events; monitor the touch for an action event associated with the set of action actions; and when associated with the set of action actions, a If the action event is not the same as the command. 52. As in the operation of claim 51, the action set depends on the touch sign, the touch position, the open application, the madness and/or the application mode. 128630.doc 200847001 5 3 · The operation of claim 5, wherein the touch or proximity touch is detected via a touch device such as a touchpad, a touch screen or a touch sensitive cover. Scheduling an action operation, comprising: monitoring a touch motion; distinguishing the touch motion between the first state and the second state; performing a first action if the touch motion is associated with the first state; and The touch motion and the second If the state is associated, a second action is performed. 55. The schematic action of claim 54 is as follows: wherein the touch motion is a slide, and the click is a tap, and the second state is a 56. A control operation comprising: an input providing a first input device and a second input device, the first device comprising a touch sensitive surface; 針對輸入事件監視該第一輸入裝置; 針對輸入事件同時監視該第二輸入裝置; 根據與4第-輸人裝置相關聯的輸人事件實行輸入操 根據與該第 入操作。 二輸入裝置相關聯的 輪入事件同時實行輸 57·如凊求項5 6之控制操作 鼠0 其中該第一輸入裝置係一滑 5 8 · —種控制操作,其包含: 128630.doc 200847001 提供一輸入功能清單,該輸入功能具有連結至該等命 令之命令及示意動作事件,料命令係關於該輸入功 能; 將輸入功能指派給弦;以及 當辨識該弦時將該輸入功能連結至弦。 59. ® 60. 61. 62. 63,Monitoring the first input device for an input event; simultaneously monitoring the second input device for an input event; and performing an input operation and the first input operation based on an input event associated with the 4th-input device. The rounding event associated with the two input devices simultaneously performs the input 57. The control operation of the request item 5 6 is performed. The first input device is a sliding control device, which includes: 128630.doc 200847001 Provided An input function list having commands and gesture events linked to the commands, the command being related to the input function; assigning the input function to the string; and coupling the input function to the string when the string is recognized. 59. ® 60. 61. 62. 63, 64. 65. 66. 如請求項此_操作,其巾該輸人功能係選自導覽操 作、檔案操作、編輯操作、檢視操作、格式化操作、工 具操作、或網頁瀏覽操作。 如請求項58之控制操作,其中該等示意動作事件係藉由 該弦實行之動作。 曰 如請求項58之控制操作,其中該等示意動作事件係選自 旋轉、作為一群組之轉換、相對於彼此之轉換、及/或敲 擊。 〆。 如請求項58之控制操作,其中以一直觀方式 動作事件連結至命彳。 …思 如請求項58之控制操作,其中將互補命令連結至互補示 意動作事件。 如請求項5 8之控制操作,其中一弦係一手指配置。 如請求項58之控制操作,其中將輸入功能指派給弦包括 從一輸入功能選單選擇一輸入功能,然後從_弦選單清 單選擇一弦。 一種控制面板,其包含: 顯示一輸入功能選單; 顯示一示意動作集,其顯示與一選定輸入功能相關聯 128630.doc 200847001 之叩令,以及指派給該等命令 顯示一弦清單, 之不思動作事件; 其中為建詈一 一各 單選擇期〜v思動作映射,—使用者從輸入功能清 坏J 2輪入功能,#您访并± 67· —種示音#&amp; 並攸該弦清單選擇一期望弦。 〜動作設計方法,其包含: 從最容易至|_ 至取難實行分級弦; 從最頻繁S曰τ 實質上”至取不頻繁分級操作頻率; 嗲等窗 b等最谷易弦與該等最頻率操作匹配,及將 與料最㈣率操作Μ 1及精細調㈣ 68·如請求項67 去’其中根據選自速度、可靠性、舒適 性、使用容易性、简s 間早性或感測區域尺寸的一或多個屬 性分級弦。 69·如請求項67之 ^ /ir 其中精細調諧係基於一或多個因 素其包括互補命令、直觀性、串擾、或組織。 7〇· —種示意動作操作,其包含: 感測一第一手指; 決定該手指之狀態; 感測一或多個額外手指; 決定額外手指之狀態;以及 根據該等第一及額外手指相對於彼此之狀態時序實施 不同輸入模式。 71·如請求項70之示意動作操作,其中若當該等額外手指處 於兩個狀態之一第一狀態時該第一手指處於兩個狀態之 128630.doc 200847001 否則實施一第二 一第一狀態,則實施—第一輸入模式 輸入模式。 包括移動或固定,以々〇 —予為之該狀您 存在或不存在八胃等額外手指之該狀態包括 -指向輪八模C該第—手指正在移動則實施 手指固^,當該等額外手指存在時該第一 只施一拖氧模式。 布64. 65. 66. If the request is for this operation, the input function is selected from navigation operations, file operations, editing operations, viewing operations, formatting operations, tool operations, or web browsing operations. The control operation of claim 58, wherein the gestures are performed by the string. The control operation of claim 58, wherein the gesture events are selected from the group consisting of rotation, conversion as a group, conversion relative to each other, and/or a hit. Hey. As in the control operation of claim 58, wherein the action event is linked to the fate in an intuitive manner. ...the control operation of claim 58, wherein the complementary command is linked to the complementary gesture event. As in the control operation of claim 58, the one string is configured with one finger. The control operation of claim 58, wherein assigning the input function to the chord includes selecting an input function from an input function menu and then selecting a chord from the _chord menu list. A control panel comprising: displaying an input function menu; displaying a set of gestures that display a command associated with a selected input function 128630.doc 200847001, and assigning a command to the command to display a list of strings, without thinking Action events; among them, for the construction of each single selection period ~ v thinking action mapping, - the user clears the J 2 round-in function from the input function, #你访和±67·—种示#&amp; The string list selects a desired string. ~ Action design method, which includes: From the easiest to |_ to the difficulty of implementing the grading string; from the most frequent S曰τ substantially "to the infrequent grading operation frequency; 嗲 窗 window b and so on the most string and such The most frequency operation matches, and the most (four) rate operation Μ 1 and fine tune (4) 68 · as requested in item 67 ', depending on the choice of speed, reliability, comfort, ease of use, simplicity, or sense One or more attributes of the area size are graded. 69. As requested in item 67 / /ir where fine tuning is based on one or more factors including complementary commands, intuitiveness, crosstalk, or organization. Illustrating an action operation, comprising: sensing a first finger; determining a state of the finger; sensing one or more additional fingers; determining a state of the additional finger; and timing based on states of the first and additional fingers relative to each other Implementing different input modes. 71. The schematic action of claim 70, wherein the first finger is in two states when the additional finger is in one of the two states, the first state is 128630.doc 200847001 Otherwise implement a second one first state, then implement - first input mode input mode. Include moving or fixed, to 々〇 - give it the shape you have or do not exist in the eight fingers and other additional fingers including - Pointing to the eight-mode C. The first finger is moving to implement the finger fixing, and when the additional fingers are present, the first one is applied to the oxygen mode. 128630.doc 10.128630.doc 10.
TW097103534A 2007-01-31 2008-01-30 Gesturing with a multipoint sensing device TW200847001A (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/700,636 US9292111B2 (en) 1998-01-26 2007-01-31 Gesturing with a multipoint sensing device

Publications (1)

Publication Number Publication Date
TW200847001A true TW200847001A (en) 2008-12-01

Family

ID=39247182

Family Applications (2)

Application Number Title Priority Date Filing Date
TW097103534A TW200847001A (en) 2007-01-31 2008-01-30 Gesturing with a multipoint sensing device
TW097202031U TWM350737U (en) 2007-01-31 2008-01-30 Electronic system

Family Applications After (1)

Application Number Title Priority Date Filing Date
TW097202031U TWM350737U (en) 2007-01-31 2008-01-30 Electronic system

Country Status (9)

Country Link
US (1) US9292111B2 (en)
EP (1) EP2115560A2 (en)
JP (1) JP2010517197A (en)
CN (1) CN101636711A (en)
AU (2) AU2008210864A1 (en)
DE (1) DE202008001338U1 (en)
HK (1) HK1109023A2 (en)
TW (2) TW200847001A (en)
WO (1) WO2008094791A2 (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101943990A (en) * 2009-07-03 2011-01-12 索尼公司 Operating control device, method of controlling operation thereof and computer program
TWI395124B (en) * 2009-04-02 2013-05-01 Mstar Semiconductor Inc Digitizing apparatus, digital converting method and capacitive touch panel apparatus
TWI413922B (en) * 2010-04-23 2013-11-01 Primax Electronics Ltd Control method for touchpad and touch device using the same
TWI455002B (en) * 2011-02-01 2014-10-01 Edamak Corp Device and method for proximity gesture detection
CN104106030A (en) * 2011-12-22 2014-10-15 纳米技术方案公司 Switched-electrode capacitive-measurement device for touch-sensitive and contactless interfaces
TWI469019B (en) * 2009-04-30 2015-01-11 Chimei Innolux Corp Power saving apparatus for capacitive touch screen and power saving method for same and capacitive touch screen employing power saving apparatus
TWI470532B (en) * 2011-06-09 2015-01-21 Inventec Corp Electronic device and control method thereof
TWI470537B (en) * 2010-12-20 2015-01-21 Apple Inc Event recognition method, related electronic device and computer readable storage medium
TWI474226B (en) * 2008-12-17 2015-02-21 Htc Corp Portable communication device and method for adjusting a plurality of touch signals thereof
TWI480797B (en) * 2013-08-14 2015-04-11 Pegatron Corp Touch method and electronic apparatus
TWI488487B (en) * 2011-10-18 2015-06-11 Acer Inc Method for adjusting video compression using gesture
TWI493407B (en) * 2009-11-09 2015-07-21 Elan Microelectronics Corp Multi - function touchpad remote control and its control method
TWI494791B (en) * 2009-11-06 2015-08-01 Au Optronics Corp Method of determining gestures for touch device
US9134897B2 (en) 2010-04-26 2015-09-15 Via Technologies, Inc. Electronic system and method for operating touch screen thereof
US9377890B2 (en) 2009-05-11 2016-06-28 Au Optronics Corp. Multi-touch method for resistive touch panel
US10402144B2 (en) 2017-05-16 2019-09-03 Wistron Corporation Portable electronic device and operation method thereof

Families Citing this family (863)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7760187B2 (en) * 2004-07-30 2010-07-20 Apple Inc. Visual expander
US7834855B2 (en) 2004-08-25 2010-11-16 Apple Inc. Wide touchpad on a portable computer
US9239673B2 (en) 1998-01-26 2016-01-19 Apple Inc. Gesturing with a multipoint sensing device
US9292111B2 (en) 1998-01-26 2016-03-22 Apple Inc. Gesturing with a multipoint sensing device
US20060033724A1 (en) * 2004-07-30 2006-02-16 Apple Computer, Inc. Virtual input device placement on a touch screen user interface
US7614008B2 (en) 2004-07-30 2009-11-03 Apple Inc. Operation of a computer with touch screen interface
US8479122B2 (en) * 2004-07-30 2013-07-02 Apple Inc. Gestures for touch sensitive input devices
US6610917B2 (en) * 1998-05-15 2003-08-26 Lester F. Ludwig Activity indication, external source, and processing loop provisions for driven vibrating-element environments
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
US7093201B2 (en) * 2001-09-06 2006-08-15 Danger, Inc. Loop menu navigation apparatus and method
US8095879B2 (en) * 2002-12-10 2012-01-10 Neonode Inc. User interface for mobile handheld computer unit
US7333092B2 (en) 2002-02-25 2008-02-19 Apple Computer, Inc. Touch pad for handheld device
JP4061274B2 (en) * 2002-03-08 2008-03-12 レベレイションズ・イン・デザイン・リミテッド・パートナーシップ Electrical device controller
US7958455B2 (en) * 2002-08-01 2011-06-07 Apple Inc. Mode activated scrolling
US8773378B2 (en) 2010-10-01 2014-07-08 Z124 Smartpad split screen
US9552141B2 (en) 2004-06-21 2017-01-24 Apple Inc. Methods and apparatuses for operating a data processing system
US7653883B2 (en) 2004-07-30 2010-01-26 Apple Inc. Proximity detector in handheld device
US8381135B2 (en) 2004-07-30 2013-02-19 Apple Inc. Proximity detector in handheld device
US7561146B1 (en) 2004-08-25 2009-07-14 Apple Inc. Method and apparatus to reject accidental contact on a touchpad
US7761814B2 (en) * 2004-09-13 2010-07-20 Microsoft Corporation Flick gesture
US8225231B2 (en) 2005-08-30 2012-07-17 Microsoft Corporation Aggregation of PC settings
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US8018440B2 (en) 2005-12-30 2011-09-13 Microsoft Corporation Unintentional touch rejection
US20070152983A1 (en) 2005-12-30 2007-07-05 Apple Computer, Inc. Touch pad with symbols based on mode
US8139028B2 (en) * 2006-02-01 2012-03-20 Synaptics Incorporated Proximity sensor and method for indicating extended interface results
US8370383B2 (en) 2006-02-08 2013-02-05 Oblong Industries, Inc. Multi-process interactive systems and methods
US8683362B2 (en) * 2008-05-23 2014-03-25 Qualcomm Incorporated Card metaphor for activities in a computing device
US8296684B2 (en) 2008-05-23 2012-10-23 Hewlett-Packard Development Company, L.P. Navigating among activities in a computing device
US8022935B2 (en) 2006-07-06 2011-09-20 Apple Inc. Capacitance sensing electrode with integrated I/O mechanism
US8564544B2 (en) 2006-09-06 2013-10-22 Apple Inc. Touch screen device, method, and graphical user interface for customizing display of content category icons
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8736557B2 (en) 2006-09-11 2014-05-27 Apple Inc. Electronic device with image based browsers
US8564543B2 (en) * 2006-09-11 2013-10-22 Apple Inc. Media player with imaged based browsing
US7581186B2 (en) * 2006-09-11 2009-08-25 Apple Inc. Media manager with integrated browsers
US8274479B2 (en) 2006-10-11 2012-09-25 Apple Inc. Gimballed scroll wheel
US8570278B2 (en) 2006-10-26 2013-10-29 Apple Inc. Portable multifunction device, method, and graphical user interface for adjusting an insertion point marker
US7856605B2 (en) * 2006-10-26 2010-12-21 Apple Inc. Method, system, and graphical user interface for positioning an insertion marker in a touch screen display
US8607167B2 (en) * 2007-01-07 2013-12-10 Apple Inc. Portable multifunction device, method, and graphical user interface for providing maps and directions
US7844915B2 (en) 2007-01-07 2010-11-30 Apple Inc. Application programming interfaces for scrolling operations
US9001047B2 (en) 2007-01-07 2015-04-07 Apple Inc. Modal change based on orientation of a portable multifunction device
US20080168402A1 (en) 2007-01-07 2008-07-10 Christopher Blumenberg Application Programming Interfaces for Gesture Operations
US20080168478A1 (en) 2007-01-07 2008-07-10 Andrew Platzer Application Programming Interfaces for Scrolling
US8607144B2 (en) * 2007-01-08 2013-12-10 Apple Inc. Monitor configuration for media device
US8612857B2 (en) * 2007-01-08 2013-12-17 Apple Inc. Monitor configuration for media device
KR100894146B1 (en) * 2007-02-03 2009-04-22 엘지전자 주식회사 Mobile communication device and control method thereof
KR101426718B1 (en) * 2007-02-15 2014-08-05 삼성전자주식회사 Apparatus and method for displaying of information according to touch event in a portable terminal
US20080229255A1 (en) * 2007-03-15 2008-09-18 Nokia Corporation Apparatus, method and system for gesture detection
KR101380004B1 (en) * 2007-03-23 2014-04-02 엘지전자 주식회사 Electronic Device and Method of executing for Application Using the Same
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
EP2150893A4 (en) * 2007-04-24 2012-08-22 Oblong Ind Inc Proteins, pools, and slawx in processing environments
US8201096B2 (en) * 2007-06-09 2012-06-12 Apple Inc. Browsing or searching user interfaces and other aspects
US8185839B2 (en) * 2007-06-09 2012-05-22 Apple Inc. Browsing or searching user interfaces and other aspects
US9933937B2 (en) 2007-06-20 2018-04-03 Apple Inc. Portable multifunction device, method, and graphical user interface for playing online videos
US8350815B2 (en) * 2007-06-20 2013-01-08 Sony Mobile Communications Portable communication device including touch input with scrolling function
US8302033B2 (en) 2007-06-22 2012-10-30 Apple Inc. Touch screen device, method, and graphical user interface for providing maps, directions, and location-based information
US8171432B2 (en) * 2008-01-06 2012-05-01 Apple Inc. Touch screen device, method, and graphical user interface for displaying and selecting application options
US8127254B2 (en) * 2007-06-29 2012-02-28 Nokia Corporation Unlocking a touch screen device
KR20090029138A (en) * 2007-09-17 2009-03-20 삼성전자주식회사 The method of inputting user command by gesture and the multimedia apparatus thereof
US9965067B2 (en) 2007-09-19 2018-05-08 T1V, Inc. Multimedia, multiuser system and associated methods
US9953392B2 (en) 2007-09-19 2018-04-24 T1V, Inc. Multimedia system and associated methods
US20130342489A1 (en) * 2008-08-13 2013-12-26 Michael R. Feldman Multimedia, multiuser system and associated methods
US7486386B1 (en) 2007-09-21 2009-02-03 Silison Laboratories Inc. Optical reflectance proximity sensor
JP2009093291A (en) * 2007-10-04 2009-04-30 Toshiba Corp Gesture determination apparatus and method
US20090100383A1 (en) * 2007-10-16 2009-04-16 Microsoft Corporation Predictive gesturing in graphical user interface
US9274698B2 (en) * 2007-10-26 2016-03-01 Blackberry Limited Electronic device and method of controlling same
US9171454B2 (en) * 2007-11-14 2015-10-27 Microsoft Technology Licensing, Llc Magic wand
US20090125848A1 (en) * 2007-11-14 2009-05-14 Susann Marie Keohane Touch surface-sensitive edit system
US8839123B2 (en) * 2007-11-19 2014-09-16 Red Hat, Inc. Generating a visual user interface
KR101413473B1 (en) * 2007-11-20 2014-07-01 엘지전자 주식회사 Mobile terminal and its key input method
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
KR101456570B1 (en) * 2007-12-21 2014-10-31 엘지전자 주식회사 Mobile terminal having digital equalizer and controlling method using the same
US8610671B2 (en) 2007-12-27 2013-12-17 Apple Inc. Insertion marker placement on touch sensitive display
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8230360B2 (en) * 2008-01-04 2012-07-24 Apple Inc. User interface for selection from media collection
US20090174679A1 (en) * 2008-01-04 2009-07-09 Wayne Carl Westerman Selective Rejection of Touch Contacts in an Edge Region of a Touch Surface
US8327272B2 (en) 2008-01-06 2012-12-04 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US8232973B2 (en) 2008-01-09 2012-07-31 Apple Inc. Method, device, and graphical user interface providing word recommendations for text input
KR20090077480A (en) * 2008-01-11 2009-07-15 삼성전자주식회사 Method for providing ui to display operation guide and multimedia apparatus thereof
US8555193B2 (en) * 2008-01-17 2013-10-08 Google Inc. System for intelligent automated layout and management of interactive windows
US20090187842A1 (en) * 2008-01-22 2009-07-23 3Dlabs Inc., Ltd. Drag and Drop User Interface for Portable Electronic Devices with Touch Sensitive Screens
KR101224588B1 (en) * 2008-01-24 2013-01-22 삼성전자주식회사 Method for providing UI to detect a multi-point stroke and multimedia apparatus thereof
KR20090085470A (en) * 2008-02-04 2009-08-07 삼성전자주식회사 A method for providing ui to detecting the plural of touch types at items or a background
US8446373B2 (en) * 2008-02-08 2013-05-21 Synaptics Incorporated Method and apparatus for extended adjustment based on relative positioning of multiple objects contemporaneously in a sensing region
US8555207B2 (en) 2008-02-27 2013-10-08 Qualcomm Incorporated Enhanced input using recognized gestures
US8717305B2 (en) 2008-03-04 2014-05-06 Apple Inc. Touch event model for web pages
US8645827B2 (en) 2008-03-04 2014-02-04 Apple Inc. Touch event model
US8201109B2 (en) 2008-03-04 2012-06-12 Apple Inc. Methods and graphical user interfaces for editing on a portable multifunction device
US8650507B2 (en) * 2008-03-04 2014-02-11 Apple Inc. Selecting of text using gestures
US9772689B2 (en) * 2008-03-04 2017-09-26 Qualcomm Incorporated Enhanced gesture-based image manipulation
US8416196B2 (en) 2008-03-04 2013-04-09 Apple Inc. Touch event model programming interface
US9513704B2 (en) * 2008-03-12 2016-12-06 Immersion Corporation Haptically enabled user interface
US20090243998A1 (en) * 2008-03-28 2009-10-01 Nokia Corporation Apparatus, method and computer program product for providing an input gesture indicator
WO2009121227A1 (en) * 2008-04-03 2009-10-08 Dong Li Method and apparatus for operating multi-object touch handheld device with touch sensitive display
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
DE102008017716A1 (en) * 2008-04-07 2009-10-08 Volkswagen Ag Displaying and operating device for vehicle, has display device, touch-sensitive position detection unit coupled with display device, and controller which is coupled with display device and position detection unit
US8335996B2 (en) * 2008-04-10 2012-12-18 Perceptive Pixel Inc. Methods of interfacing with multi-input devices and multi-input display systems employing interfacing techniques
US9740293B2 (en) 2009-04-02 2017-08-22 Oblong Industries, Inc. Operating environment with gestural control and multiple client devices, displays, and users
US10642364B2 (en) * 2009-04-02 2020-05-05 Oblong Industries, Inc. Processing tracking and recognition data in gestural recognition systems
US20090295746A1 (en) * 2008-04-29 2009-12-03 Davidson Philip L Event registration and dispatch system and method for multi-point controls
US20100177053A2 (en) * 2008-05-09 2010-07-15 Taizo Yasutake Method and apparatus for control of multiple degrees of freedom of a display
JP2009276829A (en) * 2008-05-12 2009-11-26 Nintendo Co Ltd Information processing program and information processor
US8952894B2 (en) * 2008-05-12 2015-02-10 Microsoft Technology Licensing, Llc Computer vision-based multi-touch sensing using infrared lasers
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US8174503B2 (en) 2008-05-17 2012-05-08 David H. Cain Touch-based authentication of a mobile device through user generated pattern creation
KR100914438B1 (en) * 2008-05-20 2009-08-28 엘지전자 주식회사 Electronic device with touch device and method of executing functions thereof
JP5164675B2 (en) * 2008-06-04 2013-03-21 キヤノン株式会社 User interface control method, information processing apparatus, and program
US8514251B2 (en) * 2008-06-23 2013-08-20 Qualcomm Incorporated Enhanced character input using recognized gestures
US20090327974A1 (en) * 2008-06-26 2009-12-31 Microsoft Corporation User interface for gestural control
US8504946B2 (en) * 2008-06-27 2013-08-06 Apple Inc. Portable device, method, and graphical user interface for automatically scrolling to display the top of an electronic document
US8446372B2 (en) * 2008-07-09 2013-05-21 Lenovo (Singapore) Pte. Ltd. Apparatus, system, and method for automated touchpad adjustments
US8159455B2 (en) * 2008-07-18 2012-04-17 Apple Inc. Methods and apparatus for processing combinations of kinematical inputs
US8390577B2 (en) * 2008-07-25 2013-03-05 Intuilab Continuous recognition of multi-touch gestures
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
JP4957750B2 (en) * 2008-07-31 2012-06-20 ソニー株式会社 Information processing apparatus and method, and program
US8847739B2 (en) * 2008-08-04 2014-09-30 Microsoft Corporation Fusing RFID and vision for surface object tracking
WO2010016065A1 (en) * 2008-08-08 2010-02-11 Moonsun Io Ltd. Method and device of stroke based user input
US8604364B2 (en) * 2008-08-15 2013-12-10 Lester F. Ludwig Sensors, algorithms and applications for a high dimensional touchpad
DE102008051051A1 (en) * 2008-09-03 2010-03-04 Volkswagen Ag Method and device for displaying information in a vehicle
JP5599400B2 (en) * 2008-09-04 2014-10-01 エクストリーム リアリティー エルティーディー. Method system and software for providing an image sensor based human machine interface
JPWO2010029619A1 (en) * 2008-09-10 2012-02-02 富士通東芝モバイルコミュニケーションズ株式会社 Mobile device
US20100070931A1 (en) * 2008-09-15 2010-03-18 Sony Ericsson Mobile Communications Ab Method and apparatus for selecting an object
US8769427B2 (en) * 2008-09-19 2014-07-01 Google Inc. Quick gesture input
FR2936326B1 (en) * 2008-09-22 2011-04-29 Stantum DEVICE FOR THE CONTROL OF ELECTRONIC APPARATUS BY HANDLING GRAPHIC OBJECTS ON A MULTICONTACT TOUCH SCREEN
KR20110066950A (en) * 2008-09-24 2011-06-17 코닌클리케 필립스 일렉트로닉스 엔.브이. A user interface for a multi-point touch sensitive device
US20100073303A1 (en) * 2008-09-24 2010-03-25 Compal Electronics, Inc. Method of operating a user interface
US9223430B2 (en) * 2008-09-26 2015-12-29 Hewlett-Packard Development Company Distributing touch data
US20100083108A1 (en) * 2008-09-26 2010-04-01 Research In Motion Limited Touch-screen device having soft escape key
US8957835B2 (en) * 2008-09-30 2015-02-17 Apple Inc. Head-mounted display apparatus for retaining a portable electronic device with display
US8683390B2 (en) * 2008-10-01 2014-03-25 Microsoft Corporation Manipulation of objects on multi-touch user interface
US20100087169A1 (en) * 2008-10-02 2010-04-08 Microsoft Corporation Threading together messages with multiple common participants
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
KR101586627B1 (en) * 2008-10-06 2016-01-19 삼성전자주식회사 A method for controlling of list with multi touch and apparatus thereof
KR101503835B1 (en) * 2008-10-13 2015-03-18 삼성전자주식회사 Apparatus and method for object management using multi-touch
US8446389B2 (en) * 2008-10-15 2013-05-21 Lenovo (Singapore) Pte. Ltd Techniques for creating a virtual touchscreen
US8174504B2 (en) * 2008-10-21 2012-05-08 Synaptics Incorporated Input device and method for adjusting a parameter of an electronic system
US8385952B2 (en) * 2008-10-23 2013-02-26 Microsoft Corporation Mobile communications device user interface
US8411046B2 (en) 2008-10-23 2013-04-02 Microsoft Corporation Column organization of content
US20100105441A1 (en) * 2008-10-23 2010-04-29 Chad Aron Voss Display Size of Representations of Content
US20100107100A1 (en) * 2008-10-23 2010-04-29 Schneekloth Jason S Mobile Device Style Abstraction
US20100105424A1 (en) * 2008-10-23 2010-04-29 Smuga Michael A Mobile Communications Device User Interface
US8525776B2 (en) * 2008-10-27 2013-09-03 Lenovo (Singapore) Pte. Ltd Techniques for controlling operation of a device with a virtual touchscreen
JP5684136B2 (en) * 2008-10-28 2015-03-11 サーク・コーポレーション Multi-contact area rotation gesture recognition method
JP5516412B2 (en) * 2008-10-30 2014-06-11 日本電気株式会社 Portable terminal device, data operation processing method, and data operation processing program
KR101569176B1 (en) * 2008-10-30 2015-11-20 삼성전자주식회사 Method and Apparatus for executing an object
KR101019335B1 (en) * 2008-11-11 2011-03-07 주식회사 팬택 Method and system for controlling application of mobile terminal using gesture
US8502785B2 (en) * 2008-11-12 2013-08-06 Apple Inc. Generating gestures tailored to a hand resting on a surface
KR101027566B1 (en) * 2008-11-17 2011-04-06 (주)메디슨 Ultrasonic diagnostic apparatus and method for generating commands in ultrasonic diagnostic apparatus
WO2010062263A1 (en) * 2008-11-28 2010-06-03 Creative Technology Ltd Apparatus and method for controlling a sound reproduction apparatus
SE533704C2 (en) 2008-12-05 2010-12-07 Flatfrog Lab Ab Touch sensitive apparatus and method for operating the same
US8294047B2 (en) 2008-12-08 2012-10-23 Apple Inc. Selective input signal rejection and modification
US8660300B2 (en) * 2008-12-12 2014-02-25 Silicon Laboratories Inc. Apparatus and method for optical gesture recognition
WO2010071630A1 (en) * 2008-12-15 2010-06-24 Hewlett-Packard Development Company, L.P. Gesture based edit mode
JP2010157039A (en) * 2008-12-26 2010-07-15 Toshiba Corp Electronic equipment and input control method
JP2010157038A (en) * 2008-12-26 2010-07-15 Toshiba Corp Electronic apparatus and input control method
KR101593727B1 (en) * 2008-12-29 2016-02-15 휴렛-팩커드 디벨롭먼트 컴퍼니, 엘.피. Gesture detection zones
KR101021857B1 (en) * 2008-12-30 2011-03-17 삼성전자주식회사 Apparatus and method for inputing control signal using dual touch sensor
US8957865B2 (en) * 2009-01-05 2015-02-17 Apple Inc. Device, method, and graphical user interface for manipulating a user interface object
US10019081B2 (en) * 2009-01-15 2018-07-10 International Business Machines Corporation Functionality switching in pointer input devices
US7870496B1 (en) * 2009-01-29 2011-01-11 Jahanzeb Ahmed Sherwani System using touchscreen user interface of a mobile device to remotely control a host computer
US20140160030A1 (en) * 2009-02-09 2014-06-12 Cypress Semiconductor Corporation Sensor system and method for mapping and creating gestures
TWM361674U (en) * 2009-02-19 2009-07-21 Sentelic Corp Touch control module
JP4775459B2 (en) * 2009-02-27 2011-09-21 株式会社デンソー Electronic equipment and information processing system
US9684521B2 (en) 2010-01-26 2017-06-20 Apple Inc. Systems having discrete and continuous gesture recognizers
US8589374B2 (en) 2009-03-16 2013-11-19 Apple Inc. Multifunction device with integrated search and application selection
US10705701B2 (en) 2009-03-16 2020-07-07 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US8285499B2 (en) 2009-03-16 2012-10-09 Apple Inc. Event recognition
US8566045B2 (en) 2009-03-16 2013-10-22 Apple Inc. Event recognition
US8566044B2 (en) 2009-03-16 2013-10-22 Apple Inc. Event recognition
US9311112B2 (en) * 2009-03-16 2016-04-12 Apple Inc. Event recognition
US8510665B2 (en) * 2009-03-16 2013-08-13 Apple Inc. Methods and graphical user interfaces for editing on a multifunction device with a touch screen display
US8238876B2 (en) 2009-03-30 2012-08-07 Microsoft Corporation Notifications
US8355698B2 (en) * 2009-03-30 2013-01-15 Microsoft Corporation Unlock screen
US8175653B2 (en) 2009-03-30 2012-05-08 Microsoft Corporation Chromeless user interface
US10824238B2 (en) 2009-04-02 2020-11-03 Oblong Industries, Inc. Operating environment with gestural control and multiple client devices, displays, and users
KR101593598B1 (en) * 2009-04-03 2016-02-12 삼성전자주식회사 Method for activating function of portable terminal using user gesture in portable terminal
US9258402B2 (en) * 2009-04-14 2016-02-09 Qualcomm Incorporated System and method for controlling mobile devices
KR101537706B1 (en) * 2009-04-16 2015-07-20 엘지전자 주식회사 Mobile terminal and control method thereof
KR101588730B1 (en) * 2009-04-21 2016-01-26 엘지전자 주식회사 Mobile terminal and method for communicating using instant messaging service thereof
TWI497357B (en) * 2009-04-23 2015-08-21 Waltop Int Corp Multi-touch pad control method
US8633904B2 (en) 2009-04-24 2014-01-21 Cypress Semiconductor Corporation Touch identification for multi-touch technology
US8355007B2 (en) 2009-05-11 2013-01-15 Adobe Systems Incorporated Methods for use with multi-touch displays for determining when a touch is processed as a mouse event
WO2010131122A2 (en) * 2009-05-13 2010-11-18 France Telecom User interface to provide enhanced control of an application program
US8352884B2 (en) * 2009-05-21 2013-01-08 Sony Computer Entertainment Inc. Dynamic reconfiguration of GUI display decomposition based on predictive model
US8269736B2 (en) * 2009-05-22 2012-09-18 Microsoft Corporation Drop target gestures
KR101597553B1 (en) * 2009-05-25 2016-02-25 엘지전자 주식회사 Function execution method and apparatus thereof
KR101564222B1 (en) * 2009-05-26 2015-11-06 삼성전자주식회사 Apparatus and method for unlocking a locking mode of portable terminal
US8581856B2 (en) * 2009-05-27 2013-11-12 Microsoft Corporation Touch sensitive display apparatus using sensor input
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
US20110087974A1 (en) * 2009-05-28 2011-04-14 Kulas Charles J User interface controls including capturing user mood in response to a user cue
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US20100309140A1 (en) * 2009-06-05 2010-12-09 Microsoft Corporation Controlling touch input modes
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10255566B2 (en) 2011-06-03 2019-04-09 Apple Inc. Generating and processing task items that represent tasks to perform
US8493344B2 (en) 2009-06-07 2013-07-23 Apple Inc. Devices, methods, and graphical user interfaces for accessibility using a touch-sensitive surface
US8464182B2 (en) * 2009-06-07 2013-06-11 Apple Inc. Device, method, and graphical user interface for providing maps, directions, and location-based information
US20120327009A1 (en) * 2009-06-07 2012-12-27 Apple Inc. Devices, methods, and graphical user interfaces for accessibility using a touch-sensitive surface
CA2761191C (en) * 2009-06-19 2016-09-20 Research In Motion Limited Portable electronic device with face touch detection
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US9182854B2 (en) * 2009-07-08 2015-11-10 Microsoft Technology Licensing, Llc System and method for multi-touch interactions with a touch sensitive screen
US9710097B2 (en) 2009-07-10 2017-07-18 Adobe Systems Incorporated Methods and apparatus for natural media painting using touch-and-stylus combination gestures
JP2011028366A (en) * 2009-07-22 2011-02-10 Sony Corp Operation control device and operation control method
US8428368B2 (en) * 2009-07-31 2013-04-23 Echostar Technologies L.L.C. Systems and methods for hand gesture control of an electronic device
US9563350B2 (en) * 2009-08-11 2017-02-07 Lg Electronics Inc. Mobile terminal and method for controlling the same
KR100984817B1 (en) * 2009-08-19 2010-10-01 주식회사 컴퍼니원헌드레드 User interface method using touch screen of mobile communication terminal
EP2472374B1 (en) * 2009-08-24 2019-03-20 Samsung Electronics Co., Ltd. Method for providing a ui using motions
JP5482023B2 (en) * 2009-08-27 2014-04-23 ソニー株式会社 Information processing apparatus, information processing method, and program
CN101655771B (en) * 2009-09-07 2011-07-20 上海合合信息科技发展有限公司 Method and system for inputting multi-contact characters
JP2011059820A (en) * 2009-09-07 2011-03-24 Sony Corp Information processing apparatus, information processing method and program
US9176962B2 (en) 2009-09-07 2015-11-03 Apple Inc. Digital media asset browsing with audio cues
US20110057886A1 (en) * 2009-09-10 2011-03-10 Oliver Ng Dynamic sizing of identifier on a touch-sensitive display
KR101629645B1 (en) * 2009-09-18 2016-06-21 엘지전자 주식회사 Mobile Terminal and Operation method thereof
WO2011037558A1 (en) 2009-09-22 2011-03-31 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US20120182296A1 (en) * 2009-09-23 2012-07-19 Han Dingnan Method and interface for man-machine interaction
US8799826B2 (en) * 2009-09-25 2014-08-05 Apple Inc. Device, method, and graphical user interface for moving a calendar entry in a calendar application
US8780069B2 (en) 2009-09-25 2014-07-15 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8766928B2 (en) * 2009-09-25 2014-07-01 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8832585B2 (en) 2009-09-25 2014-09-09 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US20110078626A1 (en) * 2009-09-28 2011-03-31 William Bachman Contextual Presentation of Digital Media Asset Collections
CN102033642B (en) * 2009-09-29 2012-10-10 联想(北京)有限公司 Method for gesture reorganization and electronic device
TW201112074A (en) * 2009-09-30 2011-04-01 Higgstec Inc Touch gesture detecting method of a touch panel
TW201112075A (en) * 2009-09-30 2011-04-01 Higgstec Inc Screen menu instruction generating method of a touch screen
DE102009043719A1 (en) * 2009-10-01 2011-04-07 Deutsche Telekom Ag Method for entering commands on a touch-sensitive surface
US9079498B2 (en) * 2009-10-05 2015-07-14 Tesla Motors, Inc. Morphing vehicle user interface
US8818624B2 (en) * 2009-10-05 2014-08-26 Tesla Motors, Inc. Adaptive soft buttons for a vehicle user interface
US8892299B2 (en) * 2009-10-05 2014-11-18 Tesla Motors, Inc. Vehicle user interface with proximity activation
JPWO2011043422A1 (en) * 2009-10-09 2013-03-04 日本電気株式会社 Mobile device, security lock control method thereof, and program
US8924893B2 (en) 2009-10-14 2014-12-30 At&T Mobility Ii Llc Locking and unlocking of an electronic device using a sloped lock track
US9424444B2 (en) 2009-10-14 2016-08-23 At&T Mobility Ii Llc Systems, apparatus, methods and computer-readable storage media for facilitating integrated messaging, contacts and social media for a selected entity
US9971807B2 (en) 2009-10-14 2018-05-15 Oblong Industries, Inc. Multi-process interactive systems and methods
US20110090155A1 (en) * 2009-10-15 2011-04-21 Qualcomm Incorporated Method, system, and computer program product combining gestural input from multiple touch screens into one gestural input
US20120200604A1 (en) * 2009-10-16 2012-08-09 Increment P Corporation Map display device, map display method and map display program
KR20110044496A (en) * 2009-10-23 2011-04-29 삼성전자주식회사 Input signal processing device for portable device and method including the same
US10101898B2 (en) * 2009-10-23 2018-10-16 Autodesk, Inc. Multi-touch graphical user interface for interacting with menus on a handheld device
TWI448960B (en) * 2009-11-04 2014-08-11 Univ Ishou Interactive navigation system
GB0919516D0 (en) * 2009-11-06 2009-12-23 Imagination Tech Ltd Touch sensitive display
US8390600B2 (en) * 2009-11-13 2013-03-05 Microsoft Corporation Interactive display system with contact geometry interface
US8436821B1 (en) * 2009-11-20 2013-05-07 Adobe Systems Incorporated System and method for developing and classifying touch gestures
US20110138284A1 (en) * 2009-12-03 2011-06-09 Microsoft Corporation Three-state touch input system
US20110148786A1 (en) * 2009-12-18 2011-06-23 Synaptics Incorporated Method and apparatus for changing operating modes
US9465532B2 (en) 2009-12-18 2016-10-11 Synaptics Incorporated Method and apparatus for operating in pointing and enhanced gesturing modes
JP4719296B1 (en) * 2009-12-25 2011-07-06 株式会社東芝 Information processing apparatus and information processing method
US8862576B2 (en) * 2010-01-06 2014-10-14 Apple Inc. Device, method, and graphical user interface for mapping directions between search results
US8621380B2 (en) 2010-01-06 2013-12-31 Apple Inc. Apparatus and method for conditionally enabling or disabling soft buttons
US8736561B2 (en) 2010-01-06 2014-05-27 Apple Inc. Device, method, and graphical user interface with content display modes and display rotation heuristics
US8786559B2 (en) * 2010-01-06 2014-07-22 Apple Inc. Device, method, and graphical user interface for manipulating tables using multi-contact gestures
US20110167350A1 (en) * 2010-01-06 2011-07-07 Apple Inc. Assist Features For Content Display Device
US8456297B2 (en) * 2010-01-06 2013-06-04 Apple Inc. Device, method, and graphical user interface for tracking movement on a map
US8698845B2 (en) * 2010-01-06 2014-04-15 Apple Inc. Device, method, and graphical user interface with interactive popup views
JP5750875B2 (en) * 2010-12-01 2015-07-22 ソニー株式会社 Information processing apparatus, information processing method, and program
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
JP5636678B2 (en) * 2010-01-19 2014-12-10 ソニー株式会社 Display control apparatus, display control method, and display control program
JP2011150414A (en) * 2010-01-19 2011-08-04 Sony Corp Information processing apparatus, method and program for determining operation input
US8677268B2 (en) 2010-01-26 2014-03-18 Apple Inc. Device, method, and graphical user interface for resizing objects
US8539385B2 (en) * 2010-01-26 2013-09-17 Apple Inc. Device, method, and graphical user interface for precise positioning of objects
US8539386B2 (en) * 2010-01-26 2013-09-17 Apple Inc. Device, method, and graphical user interface for selecting and moving objects
US8239785B2 (en) * 2010-01-27 2012-08-07 Microsoft Corporation Edge gestures
US9411504B2 (en) * 2010-01-28 2016-08-09 Microsoft Technology Licensing, Llc Copy and staple gestures
US20110185299A1 (en) * 2010-01-28 2011-07-28 Microsoft Corporation Stamp Gestures
US20110185320A1 (en) * 2010-01-28 2011-07-28 Microsoft Corporation Cross-reference Gestures
US8261213B2 (en) 2010-01-28 2012-09-04 Microsoft Corporation Brush, carbon-copy, and fill gestures
CN101763270B (en) 2010-01-28 2011-06-15 华为终端有限公司 Method for displaying and processing assembly and user equipment
US20110191719A1 (en) * 2010-02-04 2011-08-04 Microsoft Corporation Cut, Punch-Out, and Rip Gestures
US9519356B2 (en) * 2010-02-04 2016-12-13 Microsoft Technology Licensing, Llc Link gestures
US20110191704A1 (en) * 2010-02-04 2011-08-04 Microsoft Corporation Contextual multiplexing gestures
US9274682B2 (en) * 2010-02-19 2016-03-01 Microsoft Technology Licensing, Llc Off-screen gestures to create on-screen input
US8799827B2 (en) * 2010-02-19 2014-08-05 Microsoft Corporation Page manipulations using on and off-screen gestures
US9310994B2 (en) * 2010-02-19 2016-04-12 Microsoft Technology Licensing, Llc Use of bezel as an input mechanism
US9367205B2 (en) * 2010-02-19 2016-06-14 Microsoft Technolgoy Licensing, Llc Radial menus with bezel gestures
US9965165B2 (en) * 2010-02-19 2018-05-08 Microsoft Technology Licensing, Llc Multi-finger gestures
US20110209098A1 (en) * 2010-02-19 2011-08-25 Hinckley Kenneth P On and Off-Screen Gesture Combinations
US20110209080A1 (en) * 2010-02-24 2011-08-25 Htc Corporation Display method and electronic device for using the same
US20110205157A1 (en) * 2010-02-24 2011-08-25 Walline Erin K System and Method for Information Handling System Touchpad Enablement
US8707174B2 (en) * 2010-02-25 2014-04-22 Microsoft Corporation Multi-screen hold and page-flip gesture
US9454304B2 (en) 2010-02-25 2016-09-27 Microsoft Technology Licensing, Llc Multi-screen dual tap gesture
US9075522B2 (en) * 2010-02-25 2015-07-07 Microsoft Technology Licensing, Llc Multi-screen bookmark hold gesture
US8751970B2 (en) * 2010-02-25 2014-06-10 Microsoft Corporation Multi-screen synchronous slide gesture
US8539384B2 (en) * 2010-02-25 2013-09-17 Microsoft Corporation Multi-screen pinch and expand gestures
US20110209058A1 (en) * 2010-02-25 2011-08-25 Microsoft Corporation Multi-screen hold and tap gesture
US8473870B2 (en) * 2010-02-25 2013-06-25 Microsoft Corporation Multi-screen hold and drag gesture
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US8471814B2 (en) 2010-02-26 2013-06-25 Microsoft Corporation User interface control using a keyboard
TWI545468B (en) * 2010-03-04 2016-08-11 Sentelic Corp Input device
JP5306266B2 (en) * 2010-03-15 2013-10-02 キヤノン株式会社 Imaging apparatus and control method thereof
US8756522B2 (en) 2010-03-19 2014-06-17 Blackberry Limited Portable electronic device and method of controlling same
US9990062B2 (en) * 2010-03-26 2018-06-05 Nokia Technologies Oy Apparatus and method for proximity based input
JP5805974B2 (en) 2010-03-31 2015-11-10 ティーケー ホールディングス,インコーポレーテッド Steering wheel sensor
DE102011006344B4 (en) 2010-03-31 2020-03-12 Joyson Safety Systems Acquisition Llc Occupant measurement system
US8725230B2 (en) 2010-04-02 2014-05-13 Tk Holdings Inc. Steering wheel with hand sensors
US20110252357A1 (en) 2010-04-07 2011-10-13 Imran Chaudhri Device, Method, and Graphical User Interface for Managing Concurrently Open Software Applications
US9513801B2 (en) * 2010-04-07 2016-12-06 Apple Inc. Accessing electronic notifications and settings icons with gestures
US9823831B2 (en) 2010-04-07 2017-11-21 Apple Inc. Device, method, and graphical user interface for managing concurrently open software applications
US8448084B2 (en) * 2010-04-08 2013-05-21 Twitter, Inc. User interface mechanics
US8810509B2 (en) * 2010-04-27 2014-08-19 Microsoft Corporation Interfacing with a computing application using a multi-digit sensor
JP2011238125A (en) * 2010-05-12 2011-11-24 Sony Corp Image processing device, method and program
JP5533254B2 (en) * 2010-05-24 2014-06-25 アイシン・エィ・ダブリュ株式会社 Information display device, information display method, and program
US20110298720A1 (en) * 2010-06-02 2011-12-08 Rockwell Automation Technologies, Inc. System and method for the operation of a touch screen
EP2393000B1 (en) * 2010-06-04 2019-08-07 Lg Electronics Inc. Mobile terminal capable of providing multiplayer game and method of controlling operation of the mobile terminal
US9542091B2 (en) 2010-06-04 2017-01-10 Apple Inc. Device, method, and graphical user interface for navigating through a user interface using a dynamic object selection indicator
US8749499B2 (en) * 2010-06-08 2014-06-10 Sap Ag Touch screen for bridging multi and/or single touch points to applications
US10216408B2 (en) 2010-06-14 2019-02-26 Apple Inc. Devices and methods for identifying user interface objects based on view hierarchy
US8933910B2 (en) 2010-06-16 2015-01-13 Panasonic Intellectual Property Corporation Of America Information input apparatus, information input method, and program
US8487955B2 (en) * 2010-06-30 2013-07-16 Xerox Corporation Language-based color editing for mobile devices
JP5241038B2 (en) * 2010-07-01 2013-07-17 パナソニック株式会社 Electronic device, display control method, and program
US8773370B2 (en) 2010-07-13 2014-07-08 Apple Inc. Table editing systems with gesture-based insertion and deletion of columns and rows
US20120026077A1 (en) * 2010-07-28 2012-02-02 Google Inc. Mapping trackpad operations to touchscreen events
CN102346618A (en) * 2010-07-29 2012-02-08 鸿富锦精密工业(深圳)有限公司 Electronic device and data transmission method thereof
US9081494B2 (en) 2010-07-30 2015-07-14 Apple Inc. Device, method, and graphical user interface for copying formatting attributes
US9098182B2 (en) 2010-07-30 2015-08-04 Apple Inc. Device, method, and graphical user interface for copying user interface objects between content regions
JP5465135B2 (en) * 2010-08-30 2014-04-09 富士フイルム株式会社 MEDICAL INFORMATION DISPLAY DEVICE AND METHOD, AND PROGRAM
US8854318B2 (en) 2010-09-01 2014-10-07 Nokia Corporation Mode switching
CN101943995A (en) * 2010-09-01 2011-01-12 惠州Tcl移动通信有限公司 Method and device for processing display information of mobile terminal and touch screen thereof
JP5630160B2 (en) * 2010-09-07 2014-11-26 ソニー株式会社 Information processing apparatus, information processing method, and computer program
US20120056823A1 (en) * 2010-09-08 2012-03-08 Telefonaktiebolaget L M Ericsson (Publ) Gesture-Based Addressing of Devices
CN103081496B (en) * 2010-09-08 2016-12-07 瑞典爱立信有限公司 The control based on gesture of IPTV system
US9003298B2 (en) * 2010-09-21 2015-04-07 Microsoft Corporation Web page application controls
US9747270B2 (en) * 2011-01-07 2017-08-29 Microsoft Technology Licensing, Llc Natural input for spreadsheet actions
US20120075196A1 (en) * 2010-09-23 2012-03-29 Nokia Corporation Apparatus and method for user input
US8589350B1 (en) 2012-04-02 2013-11-19 Axcient, Inc. Systems, methods, and media for synthesizing views of file system backups
US9235474B1 (en) 2011-02-17 2016-01-12 Axcient, Inc. Systems and methods for maintaining a virtual failover volume of a target computing system
US10284437B2 (en) 2010-09-30 2019-05-07 Efolder, Inc. Cloud-based virtual machines and offices
US9705730B1 (en) 2013-05-07 2017-07-11 Axcient, Inc. Cloud storage using Merkle trees
US8924360B1 (en) 2010-09-30 2014-12-30 Axcient, Inc. Systems and methods for restoring a file
US8954544B2 (en) 2010-09-30 2015-02-10 Axcient, Inc. Cloud-based virtual machines and offices
KR101743632B1 (en) 2010-10-01 2017-06-07 삼성전자주식회사 Apparatus and method for turning e-book pages in portable terminal
US9678572B2 (en) 2010-10-01 2017-06-13 Samsung Electronics Co., Ltd. Apparatus and method for turning e-book pages in portable terminal
EP2437153A3 (en) * 2010-10-01 2016-10-05 Samsung Electronics Co., Ltd. Apparatus and method for turning e-book pages in portable terminal
US8749484B2 (en) 2010-10-01 2014-06-10 Z124 Multi-screen user interface with orientation based control
EP2437151B1 (en) 2010-10-01 2020-07-08 Samsung Electronics Co., Ltd. Apparatus and method for turning e-book pages in portable terminal
US20120218202A1 (en) 2010-10-01 2012-08-30 Sanjiv Sirpal Windows position control for phone applications
FR2965962B1 (en) * 2010-10-07 2014-11-21 Compagnie Ind Et Financiere Dingenierie Ingenico BRAILLE DATA ENTRY DEVICE, CORRESPONDING COMPUTER PROGRAM PRODUCT AND METHOD.
JP5648405B2 (en) * 2010-10-12 2015-01-07 コニカミノルタ株式会社 Image forming apparatus and computer program
KR101915615B1 (en) 2010-10-14 2019-01-07 삼성전자주식회사 Apparatus and method for controlling user interface based motion
JP5199325B2 (en) * 2010-10-14 2013-05-15 シャープ株式会社 Information input apparatus and image forming apparatus
CN102467315A (en) * 2010-10-29 2012-05-23 国际商业机器公司 Method and system for controlling electronic equipment with touch type signal input device
US8547354B2 (en) 2010-11-05 2013-10-01 Apple Inc. Device, method, and graphical user interface for manipulating soft keyboards
US8587547B2 (en) 2010-11-05 2013-11-19 Apple Inc. Device, method, and graphical user interface for manipulating soft keyboards
KR101522345B1 (en) 2010-11-12 2015-05-21 주식회사 케이티 Method for displaying background pictures in mobile communication apparatus and apparatus the same
US20120133579A1 (en) * 2010-11-30 2012-05-31 Microsoft Corporation Gesture recognition management
US8812973B1 (en) 2010-12-07 2014-08-19 Google Inc. Mobile device text-formatting
JP5604279B2 (en) * 2010-12-08 2014-10-08 日本システムウエア株式会社 Gesture recognition apparatus, method, program, and computer-readable medium storing the program
US9367154B2 (en) 2010-12-17 2016-06-14 Nec Corporation Mobile terminal device having function of flipping multiple screens by operations of multiple touches and display control method thereof
US9201539B2 (en) 2010-12-17 2015-12-01 Microsoft Technology Licensing, Llc Supplementing a touch input mechanism with fingerprint detection
CN106095418B (en) * 2010-12-20 2019-09-13 苹果公司 Event recognition
US20120159383A1 (en) 2010-12-20 2012-06-21 Microsoft Corporation Customization of an immersive environment
US9244606B2 (en) 2010-12-20 2016-01-26 Apple Inc. Device, method, and graphical user interface for navigation of concurrently open software applications
US20120159395A1 (en) 2010-12-20 2012-06-21 Microsoft Corporation Application-launching interface for multiple modes
US8689123B2 (en) 2010-12-23 2014-04-01 Microsoft Corporation Application reporting in an application-selectable user interface
US8612874B2 (en) 2010-12-23 2013-12-17 Microsoft Corporation Presenting an application change through a tile
KR101718893B1 (en) * 2010-12-24 2017-04-05 삼성전자주식회사 Method and apparatus for providing touch interface
US9223471B2 (en) * 2010-12-28 2015-12-29 Microsoft Technology Licensing, Llc Touch screen control
CN105704841B (en) * 2010-12-28 2019-03-08 联想(北京)有限公司 The method and electronic equipment of information are exchanged between a kind of electronic equipment
KR20120075839A (en) * 2010-12-29 2012-07-09 삼성전자주식회사 Method and apparatus for providing mouse right click function in touch screen terminal
KR101802498B1 (en) * 2010-12-31 2017-11-28 주식회사 케이티 Mobile terminal and method for searching location information using touch pattern recognition thereof
US9423951B2 (en) 2010-12-31 2016-08-23 Microsoft Technology Licensing, Llc Content-based snap point
TW201232349A (en) * 2011-01-21 2012-08-01 Novatek Microelectronics Corp Single finger gesture determination method, touch control chip, touch control system and computer system
US8842082B2 (en) 2011-01-24 2014-09-23 Apple Inc. Device, method, and graphical user interface for navigating and annotating an electronic document
US9250798B2 (en) * 2011-01-24 2016-02-02 Apple Inc. Device, method, and graphical user interface with a dynamic gesture disambiguation threshold
CN102622116A (en) * 2011-01-30 2012-08-01 联咏科技股份有限公司 Single-finger gesture judgment method, touch control sensing control chip and touch control system
CN103097989B (en) * 2011-01-31 2016-03-16 松下电器产业株式会社 Signal conditioning package and process control method
US8839156B2 (en) * 2011-02-03 2014-09-16 Disney Enterprises, Inc. Pointer tool for touch screens
US9665250B2 (en) 2011-02-07 2017-05-30 Blackberry Limited Portable electronic device and method of controlling same
KR101873787B1 (en) 2011-02-10 2018-07-03 삼성전자주식회사 Method for processing multi-touch input in touch screen terminal and device thereof
JP5782529B2 (en) * 2011-02-17 2015-09-24 ナイキ イノベイト シーブイ Selection of physical activity data and association with image data
KR20120095034A (en) * 2011-02-18 2012-08-28 삼성전자주식회사 Device and method for operating a touch pad in potable device
WO2012111060A1 (en) 2011-02-18 2012-08-23 Necカシオモバイルコミュニケーションズ株式会社 Electronic apparatus, control setting method, and program
US8782566B2 (en) 2011-02-22 2014-07-15 Cisco Technology, Inc. Using gestures to schedule and manage meetings
TW201237725A (en) * 2011-03-04 2012-09-16 Novatek Microelectronics Corp Single-finger and multi-touch gesture determination method, touch control chip, touch control system and computer system
US8797288B2 (en) * 2011-03-07 2014-08-05 Lester F. Ludwig Human user interfaces utilizing interruption of the execution of a first recognized gesture with the execution of a recognized second gesture
TWI483171B (en) * 2011-03-10 2015-05-01 Chi Mei Comm Systems Inc File management system and method of an electronic device
CN102681703A (en) * 2011-03-10 2012-09-19 联咏科技股份有限公司 Single-finger and multi-finger gesture judging method, touch induction control chip and touch system
US8760424B2 (en) * 2011-03-17 2014-06-24 Intellitact Llc Touch enhanced interface
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US9383917B2 (en) 2011-03-28 2016-07-05 Microsoft Technology Licensing, Llc Predictive tiling
JP5716502B2 (en) * 2011-04-06 2015-05-13 ソニー株式会社 Information processing apparatus, information processing method, and computer program
JP5716503B2 (en) * 2011-04-06 2015-05-13 ソニー株式会社 Information processing apparatus, information processing method, and computer program
TWI483172B (en) * 2011-04-07 2015-05-01 Chi Mei Comm Systems Inc Method and system for arranging a user interface of the electronic device
US9367229B2 (en) * 2011-04-08 2016-06-14 Adobe Systems Incorporated Touch-enabled device navigation system
US9298363B2 (en) 2011-04-11 2016-03-29 Apple Inc. Region activation for touch sensitive surface
KR20130052749A (en) * 2011-04-15 2013-05-23 엘지전자 주식회사 Touch based user interface device and methdo
KR101199618B1 (en) 2011-05-11 2012-11-08 주식회사 케이티테크 Apparatus and Method for Screen Split Displaying
US9329773B2 (en) * 2011-05-19 2016-05-03 International Business Machines Corporation Scalable gesture-based device control
JP2012243163A (en) * 2011-05-20 2012-12-10 Sony Corp Electronic device, program, and control method
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US9104307B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US8893033B2 (en) 2011-05-27 2014-11-18 Microsoft Corporation Application notifications
US20120304132A1 (en) 2011-05-27 2012-11-29 Chaitanya Dev Sareen Switching back to a previously-interacted-with application
US9032338B2 (en) 2011-05-30 2015-05-12 Apple Inc. Devices, methods, and graphical user interfaces for navigating and editing text
US8661339B2 (en) 2011-05-31 2014-02-25 Apple Inc. Devices, methods, and graphical user interfaces for document manipulation
CN103608760A (en) 2011-06-03 2014-02-26 谷歌公司 Gestures for selecting text
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US9513799B2 (en) * 2011-06-05 2016-12-06 Apple Inc. Devices, methods, and graphical user interfaces for providing control of a touch-based user interface absent physical touch capabilities
US20120311444A1 (en) * 2011-06-05 2012-12-06 Apple Inc. Portable multifunction device, method, and graphical user interface for controlling media playback using gestures
KR101810884B1 (en) * 2011-06-07 2017-12-20 삼성전자주식회사 Apparatus and method for providing web browser interface using gesture in device
JP5790203B2 (en) * 2011-06-29 2015-10-07 ソニー株式会社 Information processing apparatus, information processing method, program, and remote operation system
JP2013020289A (en) * 2011-07-07 2013-01-31 Alps Electric Co Ltd Input processor
KR101948645B1 (en) 2011-07-11 2019-02-18 삼성전자 주식회사 Method and apparatus for controlling contents using graphic object
US9582187B2 (en) 2011-07-14 2017-02-28 Microsoft Technology Licensing, Llc Dynamic context based menus
US20130019175A1 (en) * 2011-07-14 2013-01-17 Microsoft Corporation Submenus for context based menu system
CN102902469B (en) * 2011-07-25 2015-08-19 宸鸿光电科技股份有限公司 Gesture identification method and touch-control system
US8687023B2 (en) 2011-08-02 2014-04-01 Microsoft Corporation Cross-slide gesture to select and rearrange
US9417754B2 (en) 2011-08-05 2016-08-16 P4tents1, LLC User interface system, method, and computer program product
WO2013022222A2 (en) * 2011-08-05 2013-02-14 Samsung Electronics Co., Ltd. Method for controlling electronic apparatus based on motion recognition, and electronic apparatus applying the same
EP4235365A3 (en) 2011-08-05 2023-11-01 Samsung Electronics Co., Ltd. Method for controlling electronic apparatus based on voice recognition and motion recognition, and electronic apparatus applying the same
EP2557397B1 (en) * 2011-08-12 2014-02-19 Technisat Digital Gmbh Method for determining a route using a navigation system
KR20130052753A (en) * 2011-08-16 2013-05-23 삼성전자주식회사 Method of executing application using touchscreen and terminal supporting the same
US10706096B2 (en) 2011-08-18 2020-07-07 Apple Inc. Management of local and remote media items
US8806369B2 (en) 2011-08-26 2014-08-12 Apple Inc. Device, method, and graphical user interface for managing and interacting with concurrently open software applications
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
US20130057587A1 (en) 2011-09-01 2013-03-07 Microsoft Corporation Arranging tiles
JP5576841B2 (en) 2011-09-09 2014-08-20 Kddi株式会社 User interface device capable of zooming image by pressing, image zoom method and program
US10353566B2 (en) 2011-09-09 2019-07-16 Microsoft Technology Licensing, Llc Semantic zoom animations
US9557909B2 (en) 2011-09-09 2017-01-31 Microsoft Technology Licensing, Llc Semantic zoom linguistic helpers
US8922575B2 (en) 2011-09-09 2014-12-30 Microsoft Corporation Tile cache
US9244802B2 (en) 2011-09-10 2016-01-26 Microsoft Technology Licensing, Llc Resource user interface
US9146670B2 (en) 2011-09-10 2015-09-29 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
US8933952B2 (en) 2011-09-10 2015-01-13 Microsoft Corporation Pre-rendering new content for an application-selectable user interface
JP2014530417A (en) * 2011-09-13 2014-11-17 ニューマン インフィニット, インク.Newman Infinite, Inc. Apparatus, method and computer readable storage medium for operating user interface elements
EP2573666B1 (en) * 2011-09-21 2018-09-05 LG Electronics Inc. Mobile terminal and control method thereof
KR101869774B1 (en) * 2011-09-21 2018-06-22 엘지전자 주식회사 Mobile terminal and control method for mobile terminal
US8842057B2 (en) 2011-09-27 2014-09-23 Z124 Detail on triggers: transitional states
WO2013051050A1 (en) 2011-10-03 2013-04-11 古野電気株式会社 Device having touch panel, radar device, plotter device, marine network system, symbol identification method and symbol identification program
WO2013051054A1 (en) 2011-10-03 2013-04-11 古野電気株式会社 Device having touch panel, radar device, plotter device, network system for ships, information display method and information display program
US9157750B2 (en) 2011-10-03 2015-10-13 Furuno Electric Co., Ltd. Device having touch panel, radar apparatus, plotter apparatus, ship network system, information display method and information display program
WO2013051049A1 (en) 2011-10-03 2013-04-11 古野電気株式会社 Device having touch panel, radar device, plotter device, network system for use on ship, perspective-change method, and perspective-change program
EP2769291B1 (en) 2011-10-18 2021-04-28 Carnegie Mellon University Method and apparatus for classifying touch events on a touch sensitive surface
US9053083B2 (en) 2011-11-04 2015-06-09 Microsoft Technology Licensing, Llc Interaction between web gadgets and spreadsheets
US9367230B2 (en) * 2011-11-08 2016-06-14 Microsoft Technology Licensing, Llc Interaction models for indirect interaction devices
TWI597626B (en) 2011-11-08 2017-09-01 威盛電子股份有限公司 Control device of touch panel, method for determining single-point-multiple-finger gesture and touch pen for generating single-point-multiple-finger gesture
US10082950B2 (en) * 2011-11-09 2018-09-25 Joseph T. LAPP Finger-mapped character entry systems
EP2776908A4 (en) 2011-11-09 2015-07-15 Blackberry Ltd Touch-sensitive display method and apparatus
US9383858B2 (en) * 2011-11-23 2016-07-05 Guangzhou Ucweb Computer Technology Co., Ltd Method and device for executing an operation on a mobile device
CN103135755B (en) * 2011-12-02 2016-04-06 深圳泰山在线科技有限公司 Interactive system and method
US9286414B2 (en) 2011-12-02 2016-03-15 Microsoft Technology Licensing, Llc Data discovery and description service
US20130155010A1 (en) * 2011-12-14 2013-06-20 Microchip Technology Incorporated Capacitive Proximity Based Gesture Input System
US8902180B2 (en) * 2011-12-16 2014-12-02 Nokia Corporation Methods, apparatuses, and computer program products for enabling use of remote devices with pre-defined gestures
US9292094B2 (en) 2011-12-16 2016-03-22 Microsoft Technology Licensing, Llc Gesture inferred vocabulary bindings
DE102011056940A1 (en) 2011-12-22 2013-06-27 Bauhaus Universität Weimar A method of operating a multi-touch display and device having a multi-touch display
CN106648434B (en) * 2011-12-22 2020-09-04 阿里巴巴(中国)有限公司 Method and device for controlling application interface through dragging gesture
US9223472B2 (en) 2011-12-22 2015-12-29 Microsoft Technology Licensing, Llc Closing applications
US9116611B2 (en) * 2011-12-29 2015-08-25 Apple Inc. Devices, methods, and graphical user interfaces for providing multitouch inputs and hardware-based features using a single touch input
US9342181B2 (en) 2012-01-09 2016-05-17 Nvidia Corporation Touch-screen input/output device touch sensing techniques
KR101951480B1 (en) * 2012-01-09 2019-02-22 엘지전자 주식회사 Electronic Device And Method Of Controlling The Same
KR20130083195A (en) * 2012-01-12 2013-07-22 삼성전자주식회사 Method and apparatus for keyboard layout using touch
TW201329837A (en) * 2012-01-13 2013-07-16 Fih Hong Kong Ltd System and method for unlocking an electronic device
US9171099B2 (en) 2012-01-26 2015-10-27 Microsoft Technology Licensing, Llc System and method for providing calculation web services for online documents
KR20130091046A (en) * 2012-02-07 2013-08-16 김지영 The apparatus of display in idea wheel structure
US9395901B2 (en) * 2012-02-08 2016-07-19 Blackberry Limited Portable electronic device and method of controlling same
EP2626778B1 (en) 2012-02-09 2017-04-05 Sony Mobile Communications, Inc. Capacitive touch panel device
US9128605B2 (en) 2012-02-16 2015-09-08 Microsoft Technology Licensing, Llc Thumbnail-image selection of applications
KR101873413B1 (en) * 2012-02-17 2018-07-02 엘지전자 주식회사 Mobile terminal and control method for the mobile terminal
US20130215038A1 (en) * 2012-02-17 2013-08-22 Rukman Senanayake Adaptable actuated input device with integrated proximity detection
US8830193B2 (en) 2012-02-23 2014-09-09 Honeywell International Inc. Controlling views in display device with touch screen
US9977876B2 (en) 2012-02-24 2018-05-22 Perkinelmer Informatics, Inc. Systems, methods, and apparatus for drawing chemical structures using touch and gestures
KR101894567B1 (en) 2012-02-24 2018-09-03 삼성전자 주식회사 Operation Method of Lock Screen And Electronic Device supporting the same
JP2013179402A (en) * 2012-02-28 2013-09-09 Sony Corp Terminal device, information processor, display method, and display control method
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
WO2013138052A1 (en) * 2012-03-13 2013-09-19 Microsoft Corporation Web page application controls
US20130246975A1 (en) * 2012-03-15 2013-09-19 Chandar Kumar Oddiraju Gesture group selection
US8803831B1 (en) 2012-03-23 2014-08-12 Google Inc. Chording sheath for computing device
US9575652B2 (en) 2012-03-31 2017-02-21 Microsoft Technology Licensing, Llc Instantiable gesture objects
US8881269B2 (en) 2012-03-31 2014-11-04 Apple Inc. Device, method, and graphical user interface for integrating recognition of handwriting gestures with a screen reader
JP5663519B2 (en) * 2012-04-10 2015-02-04 京セラドキュメントソリューションズ株式会社 Display input device and image forming apparatus
JP5790578B2 (en) * 2012-04-10 2015-10-07 株式会社デンソー Display system, display device, and operation device
JP5645869B2 (en) * 2012-04-13 2014-12-24 京セラドキュメントソリューションズ株式会社 Display input device and image forming apparatus
WO2013154720A1 (en) 2012-04-13 2013-10-17 Tk Holdings Inc. Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same
CN102622182A (en) * 2012-04-16 2012-08-01 李波 Off-screen touch interactive system with nearly-joint sensing projection point coordinate indication
US20130285924A1 (en) * 2012-04-26 2013-10-31 Research In Motion Limited Method and Apparatus Pertaining to the Interpretation of Touch-Based Actions
TWI456487B (en) * 2012-04-26 2014-10-11 Acer Inc Mobile device and gesture determination method
US9772700B2 (en) * 2012-04-30 2017-09-26 Blackberry Limited Device and method for processing user input
US20130285927A1 (en) * 2012-04-30 2013-10-31 Research In Motion Limited Touchscreen keyboard with correction of previously input text
US20140040789A1 (en) * 2012-05-08 2014-02-06 Adobe Systems Incorporated Tool configuration history in a user interface
WO2013169853A1 (en) 2012-05-09 2013-11-14 Industries Llc Yknots Device, method, and graphical user interface for providing tactile feedback for operations performed in a user interface
WO2013169865A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for moving a user interface object based on an intensity of a press input
EP3264252B1 (en) 2012-05-09 2019-11-27 Apple Inc. Device, method, and graphical user interface for performing an operation in accordance with a selected mode of operation
KR101806350B1 (en) 2012-05-09 2017-12-07 애플 인크. Device, method, and graphical user interface for selecting user interface objects
WO2013169849A2 (en) 2012-05-09 2013-11-14 Industries Llc Yknots Device, method, and graphical user interface for displaying user interface objects corresponding to an application
AU2013259606B2 (en) 2012-05-09 2016-06-02 Apple Inc. Device, method, and graphical user interface for displaying additional information in response to a user contact
KR101823288B1 (en) 2012-05-09 2018-01-29 애플 인크. Device, method, and graphical user interface for transitioning between display states in response to gesture
WO2013169843A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for manipulating framed graphical objects
WO2013169875A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for displaying content associated with a corresponding affordance
WO2013169851A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for facilitating user interaction with controls in a user interface
WO2013169845A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for scrolling nested regions
WO2013169854A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for providing feedback for changing activation states of a user interface object
WO2013169842A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for selecting object within a group of objects
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US20130307788A1 (en) * 2012-05-16 2013-11-21 Motorola Solutions, Inc. Device and method for automated use of force sensing touch panels
US10168835B2 (en) 2012-05-23 2019-01-01 Flatfrog Laboratories Ab Spatial resolution in touch displays
CA2793763C (en) * 2012-05-24 2015-12-29 Research In Motion Limited Presentation of image on display screen with combination crop and rotation and with auto-resizing of crop field
CN103455243B (en) * 2012-06-04 2016-09-28 宏达国际电子股份有限公司 Adjust the method and device of screen object size
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9223489B2 (en) * 2012-06-13 2015-12-29 Adobe Systems Incorporated Method and apparatus for gesture based copying of attributes
CN102750096A (en) * 2012-06-15 2012-10-24 深圳乐投卡尔科技有限公司 Vehicle-mounted Android platform multi-point gesture control method
US9487388B2 (en) 2012-06-21 2016-11-08 Nextinput, Inc. Ruggedized MEMS force die
US8970525B1 (en) 2012-06-27 2015-03-03 Google Inc. Method and system for trackpad input error mitigation
CN102779009B (en) * 2012-06-29 2015-04-08 华为终端有限公司 Method and terminal for displaying application program interface
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
CN103513785A (en) * 2012-06-29 2014-01-15 联想(北京)有限公司 Method for processing information, electronic device and input device
US9032818B2 (en) 2012-07-05 2015-05-19 Nextinput, Inc. Microelectromechanical load sensor and methods of manufacturing the same
CN109508091A (en) * 2012-07-06 2019-03-22 原相科技股份有限公司 Input system
KR20150034255A (en) 2012-07-15 2015-04-02 애플 인크. Disambiguation of multitouch gesture recognition for 3d interaction
US9268424B2 (en) * 2012-07-18 2016-02-23 Sony Corporation Mobile client device, operation method, recording medium, and operation system
US9058103B2 (en) * 2012-07-25 2015-06-16 Facebook, Inc. Gestures for keyboard switch
US9823935B2 (en) * 2012-07-26 2017-11-21 Nvidia Corporation Techniques for latching input events to display flips
TW201405363A (en) * 2012-07-26 2014-02-01 Hon Hai Prec Ind Co Ltd Application controlling system and method
KR102101818B1 (en) * 2012-07-30 2020-04-17 삼성전자주식회사 Device and method for controlling data transfer in terminal
CN103576961A (en) * 2012-08-06 2014-02-12 闫化玉 Gesture man-machine interaction method using general symbols and variant symbols thereof
US9507513B2 (en) 2012-08-17 2016-11-29 Google Inc. Displaced double tap gesture
US9250783B2 (en) 2012-08-21 2016-02-02 Apple Inc. Toggle gesture during drag gesture
JP5928245B2 (en) * 2012-08-23 2016-06-01 カシオ計算機株式会社 Data processing apparatus and program
US9552068B2 (en) * 2012-08-27 2017-01-24 Microchip Technology Germany Gmbh Input device with hand posture control
US9558278B2 (en) 2012-09-11 2017-01-31 Apple Inc. Integrated content recommendation
US9218118B2 (en) 2012-09-11 2015-12-22 Apple Inc. Media player playlist management
US20150227221A1 (en) * 2012-09-12 2015-08-13 Toyota Jidosha Kabushiki Kaisha Mobile terminal device, on-vehicle device, and on-vehicle system
US9696223B2 (en) 2012-09-17 2017-07-04 Tk Holdings Inc. Single layer force sensor
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
US9020845B2 (en) 2012-09-25 2015-04-28 Alexander Hieronymous Marlowe System and method for enhanced shopping, preference, profile and survey data input and gathering
US9785647B1 (en) 2012-10-02 2017-10-10 Axcient, Inc. File system virtualization
US9430123B2 (en) * 2012-10-09 2016-08-30 Sap Se Triggering a refresh of displayed content on a mobile device
US11386257B2 (en) 2012-10-15 2022-07-12 Amaze Software, Inc. Efficient manipulation of surfaces in multi-dimensional space using energy agents
US10877780B2 (en) 2012-10-15 2020-12-29 Famous Industries, Inc. Visibility detection using gesture fingerprinting
US10908929B2 (en) 2012-10-15 2021-02-02 Famous Industries, Inc. Human versus bot detection using gesture fingerprinting
US9501171B1 (en) * 2012-10-15 2016-11-22 Famous Industries, Inc. Gesture fingerprinting
JP5552632B2 (en) * 2012-10-29 2014-07-16 株式会社東芝 Information processing apparatus and input method
US9058100B2 (en) * 2012-11-06 2015-06-16 Fang Li Using motion gestures to send, save, delete, and reject a message
US9852140B1 (en) 2012-11-07 2017-12-26 Axcient, Inc. Efficient file replication
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US10503359B2 (en) * 2012-11-15 2019-12-10 Quantum Interface, Llc Selection attractive interfaces, systems and apparatuses including such interfaces, methods for making and using same
US10289204B2 (en) * 2012-11-15 2019-05-14 Quantum Interface, Llc Apparatuses for controlling electrical devices and software programs and methods for making and using same
WO2014083953A1 (en) * 2012-11-27 2014-06-05 ソニー株式会社 Display device, display method, and computer program
US9235342B2 (en) 2012-11-28 2016-01-12 International Business Machines Corporation Selective sharing of displayed content in a view presented on a touchscreen of a processing system
CN103019378B (en) * 2012-12-07 2016-09-07 无锡清华信息科学与技术国家实验室物联网技术中心 A kind of mobile electronic device gesture control interaction method, device and mobile terminal
US9535583B2 (en) 2012-12-13 2017-01-03 Perkinelmer Informatics, Inc. Draw-ahead feature for chemical structure drawing applications
US9134892B2 (en) 2012-12-14 2015-09-15 Barnes & Noble College Booksellers, Llc Drag-based content selection technique for touch screen UI
US9134893B2 (en) * 2012-12-14 2015-09-15 Barnes & Noble College Booksellers, Llc Block-based content selecting technique for touch screen UI
US20140173529A1 (en) * 2012-12-14 2014-06-19 Barnesandnoble.Com Llc Circular gesture for touch sensitive ui control feature
US9659546B2 (en) * 2012-12-17 2017-05-23 Nec Display Solutions, Ltd. Control apparatus and control method to control external device depending on motion of human body
TWI475472B (en) * 2012-12-19 2015-03-01 Inventec Corp System for drawing on touch screen and method thereof
CN103902080A (en) * 2012-12-27 2014-07-02 华硕电脑股份有限公司 Touch device and touch processing method
KR102301592B1 (en) 2012-12-29 2021-09-10 애플 인크. Device, method, and graphical user interface for navigating user interface hierachies
WO2014105279A1 (en) * 2012-12-29 2014-07-03 Yknots Industries Llc Device, method, and graphical user interface for switching between user interfaces
WO2014105277A2 (en) 2012-12-29 2014-07-03 Yknots Industries Llc Device, method, and graphical user interface for moving a cursor according to a change in an appearance of a control icon with simulated three-dimensional characteristics
WO2014105278A1 (en) 2012-12-29 2014-07-03 Yknots Industries Llc Device, method, and graphical user interface for determining whether to scroll or select contents
CN108845748A (en) 2012-12-29 2018-11-20 苹果公司 For abandoning generating equipment, method and the graphic user interface of tactile output for more contact gestures
KR101958517B1 (en) 2012-12-29 2019-03-14 애플 인크. Device, method, and graphical user interface for transitioning between touch input to display output relationships
KR102064965B1 (en) 2013-01-04 2020-01-10 엘지전자 주식회사 Method for controlling using double touch jesture and the terminal thereof
WO2014113462A1 (en) 2013-01-15 2014-07-24 Cirque Corporation Multi-dimensional multi-finger search using oversampling hill climbing and descent with range
US9323343B2 (en) * 2013-01-31 2016-04-26 Panasonic Intellectual Property Corporation Of America Information processing method and information processing apparatus
CN103135929A (en) * 2013-01-31 2013-06-05 北京小米科技有限责任公司 Method and device for controlling application interface to move and terminal device
JP6232706B2 (en) * 2013-02-05 2017-11-22 コニカミノルタ株式会社 INFORMATION DISPLAY DEVICE, IMAGE FORMING DEVICE, INFORMATION DISPLAY DEVICE CONTROL METHOD, AND INFORMATION DISPLAY DEVICE CONTROL PROGRAM
US20140237422A1 (en) * 2013-02-15 2014-08-21 Flatfrog Laboratories Ab Interpretation of pressure based gesture
US20140232672A1 (en) * 2013-02-20 2014-08-21 Tencent Technology (Shenzhen) Company Limited Method and terminal for triggering application programs and application program functions
JP5870212B2 (en) * 2013-02-20 2016-02-24 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America Control method and program for portable information terminal
US9292153B1 (en) * 2013-03-07 2016-03-22 Axcient, Inc. Systems and methods for providing efficient and focused visualization of data
US9397907B1 (en) 2013-03-07 2016-07-19 Axcient, Inc. Protection status determinations for computing devices
KR102117086B1 (en) * 2013-03-08 2020-06-01 삼성디스플레이 주식회사 Terminal and method for controlling thereof
US8854361B1 (en) 2013-03-13 2014-10-07 Cambridgesoft Corporation Visually augmenting a graphical rendering of a chemical structure representation or biological sequence representation with multi-dimensional information
US9823813B2 (en) * 2013-03-15 2017-11-21 Salesforce.Com, Inc. Apparatus and methods for performing an action on a database record
US20140280329A1 (en) 2013-03-15 2014-09-18 Salesforce.Com, Inc. Systems and methods for cross-referencing feed items
US9658740B2 (en) 2013-03-15 2017-05-23 Apple Inc. Device, method, and graphical user interface for managing concurrently open software applications
CN105144071B (en) * 2013-03-15 2019-11-19 苹果公司 Method and electronic equipment for managing user interface
US9170676B2 (en) 2013-03-15 2015-10-27 Qualcomm Incorporated Enhancing touch inputs with gestures
US9477404B2 (en) 2013-03-15 2016-10-25 Apple Inc. Device, method, and graphical user interface for managing concurrently open software applications
CN104063140B (en) * 2013-03-18 2017-11-03 联想(北京)有限公司 The method and electronic equipment of selecting object
KR20140114766A (en) 2013-03-19 2014-09-29 퀵소 코 Method and device for sensing touch inputs
US9013452B2 (en) 2013-03-25 2015-04-21 Qeexo, Co. Method and system for activating different interactive functions using different types of finger contacts
US9612689B2 (en) 2015-02-02 2017-04-04 Qeexo, Co. Method and apparatus for classifying a touch event on a touchscreen as related to one of multiple function generating interaction layers and activating a function in the selected interaction layer
US9715282B2 (en) * 2013-03-29 2017-07-25 Microsoft Technology Licensing, Llc Closing, starting, and restarting applications
US9450952B2 (en) 2013-05-29 2016-09-20 Microsoft Technology Licensing, Llc Live tiles without application-code execution
US20140298219A1 (en) * 2013-03-29 2014-10-02 Microsoft Corporation Visual Selection and Grouping
US10114536B2 (en) * 2013-03-29 2018-10-30 Microsoft Technology Licensing, Llc Systems and methods for performing actions for users from a locked device
KR102087896B1 (en) * 2013-04-03 2020-03-12 삼성전자주식회사 Method and apparatus for inputting text in electronic device having touchscreen
US9589120B2 (en) * 2013-04-05 2017-03-07 Microsoft Technology Licensing, Llc Behavior based authentication for touch screen devices
WO2014168567A1 (en) 2013-04-11 2014-10-16 Flatfrog Laboratories Ab Tomographic processing for touch detection
TWI493406B (en) * 2013-04-24 2015-07-21 Acer Inc Electronic apparatus and touch detecting method thereof
CN104123089A (en) * 2013-04-27 2014-10-29 腾讯科技(深圳)有限公司 Gesture operation method and device for address bar and touch screen terminal
US10599250B2 (en) * 2013-05-06 2020-03-24 Qeexo, Co. Using finger touch types to interact with electronic devices
US9430127B2 (en) 2013-05-08 2016-08-30 Cambridgesoft Corporation Systems and methods for providing feedback cues for touch screen interface interaction with chemical and biological structure drawing applications
US9751294B2 (en) 2013-05-09 2017-09-05 Perkinelmer Informatics, Inc. Systems and methods for translating three dimensional graphic molecular models to computer aided design format
US10141930B2 (en) 2013-06-04 2018-11-27 Nvidia Corporation Three state latch
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
DE112014002747T5 (en) 2013-06-09 2016-03-03 Apple Inc. Apparatus, method and graphical user interface for enabling conversation persistence over two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
US9733716B2 (en) 2013-06-09 2017-08-15 Apple Inc. Proxy gesture recognizer
US10664652B2 (en) 2013-06-15 2020-05-26 Microsoft Technology Licensing, Llc Seamless grid and canvas integration in a spreadsheet application
CN104252302A (en) * 2013-06-26 2014-12-31 富泰华工业(深圳)有限公司 Self-adaptive image adjusting system and method
US9874978B2 (en) 2013-07-12 2018-01-23 Flatfrog Laboratories Ab Partial detect mode
US10809893B2 (en) * 2013-08-09 2020-10-20 Insyde Software Corp. System and method for re-sizing and re-positioning application windows in a touch-based computing device
US9110561B2 (en) * 2013-08-12 2015-08-18 Apple Inc. Context sensitive actions
JP5456200B1 (en) * 2013-08-13 2014-03-26 ソフトバンクモバイル株式会社 Character processing apparatus and program
US9547525B1 (en) 2013-08-21 2017-01-17 Google Inc. Drag toolbar to enter tab switching interface
CN105474112B (en) * 2013-08-21 2019-03-08 三菱电机株式会社 Program editing method
US9569004B2 (en) 2013-08-22 2017-02-14 Google Inc. Swipe toolbar to switch tabs
US9785241B2 (en) * 2013-08-26 2017-10-10 Paypal, Inc. Gesture identification
US9823728B2 (en) 2013-09-04 2017-11-21 Nvidia Corporation Method and system for reduced rate touch scanning on an electronic device
JP6223755B2 (en) * 2013-09-06 2017-11-01 株式会社東芝 Method, electronic device, and program
JP2015064697A (en) * 2013-09-24 2015-04-09 富士通株式会社 Information processing apparatus, information processing method, and information processing program
US9542004B1 (en) * 2013-09-24 2017-01-10 Amazon Technologies, Inc. Gesture-based flash
US9881592B2 (en) 2013-10-08 2018-01-30 Nvidia Corporation Hardware overlay assignment
EP3058448A4 (en) * 2013-10-18 2017-04-12 Citrix Systems Inc. Providing enhanced message management user interfaces
US9304597B2 (en) 2013-10-29 2016-04-05 Intel Corporation Gesture based human computer interaction
US9965170B2 (en) * 2013-11-11 2018-05-08 Lenovo (Singapore) Pte. Ltd. Multi-touch inputs for input interface control
TWI489355B (en) * 2013-11-13 2015-06-21 Wistron Corp Touch sensing module, touch sensing method, and computer program product
KR102206053B1 (en) * 2013-11-18 2021-01-21 삼성전자주식회사 Apparatas and method for changing a input mode according to input method in an electronic device
JP2015103161A (en) * 2013-11-27 2015-06-04 コニカミノルタ株式会社 Image formation device, display method of operation screen, and computer program
JP6062351B2 (en) * 2013-11-28 2017-01-18 京セラ株式会社 Electronics
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9507470B2 (en) 2013-12-16 2016-11-29 Nvidia Corporation Method and system for reduced power touch input detection on an electronic device using reduced scanning
US9902611B2 (en) 2014-01-13 2018-02-27 Nextinput, Inc. Miniaturized and ruggedized wafer level MEMs force sensors
WO2015108479A1 (en) 2014-01-16 2015-07-23 Flatfrog Laboratories Ab Light coupling in tir-based optical touch systems
WO2015108480A1 (en) 2014-01-16 2015-07-23 Flatfrog Laboratories Ab Improvements in tir-based optical touch systems of projection-type
JP6278262B2 (en) * 2014-03-12 2018-02-14 ヤマハ株式会社 Display control device
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
CN103870061A (en) * 2014-03-14 2014-06-18 李雪飞 Method for realizing mouse function on multi-point touch control equipment
KR102298602B1 (en) 2014-04-04 2021-09-03 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 Expandable application representation
KR20160143784A (en) 2014-04-10 2016-12-14 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 Slider cover for computing device
KR102107275B1 (en) 2014-04-10 2020-05-06 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 Collapsible shell cover for computing device
US10579252B2 (en) * 2014-04-28 2020-03-03 Ford Global Technologies, Llc Automotive touchscreen with simulated texture for the visually impaired
US11625145B2 (en) 2014-04-28 2023-04-11 Ford Global Technologies, Llc Automotive touchscreen with simulated texture for the visually impaired
DE102014208222A1 (en) * 2014-04-30 2015-11-05 Siemens Aktiengesellschaft A method of retrieving application commands, arithmetic unit, and medical imaging system
KR20150134674A (en) * 2014-05-22 2015-12-02 삼성전자주식회사 User terminal device, and Method for controlling for User terminal device, and multimedia system thereof
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10204096B2 (en) * 2014-05-30 2019-02-12 Apple Inc. Device, method, and graphical user interface for a predictive keyboard
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
EP3149728B1 (en) 2014-05-30 2019-01-16 Apple Inc. Multi-command single utterance input method
US9898162B2 (en) 2014-05-30 2018-02-20 Apple Inc. Swiping functions for messaging applications
US9971500B2 (en) 2014-06-01 2018-05-15 Apple Inc. Displaying options, assigning notification, ignoring messages, and simultaneous user interface displays in a messaging application
US9575560B2 (en) 2014-06-03 2017-02-21 Google Inc. Radar-based gesture-recognition through a wearable device
CN105224114A (en) * 2014-06-11 2016-01-06 天津富纳源创科技有限公司 Touch plate control method
WO2015199602A1 (en) 2014-06-27 2015-12-30 Flatfrog Laboratories Ab Detection of surface contamination
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
JP6399834B2 (en) * 2014-07-10 2018-10-03 キヤノン株式会社 Information processing apparatus, information processing apparatus control method, and program
US10592080B2 (en) 2014-07-31 2020-03-17 Microsoft Technology Licensing, Llc Assisted presentation of application windows
US10678412B2 (en) 2014-07-31 2020-06-09 Microsoft Technology Licensing, Llc Dynamic joint dividers for application windows
US10254942B2 (en) 2014-07-31 2019-04-09 Microsoft Technology Licensing, Llc Adaptive sizing and positioning of application windows
DE102014111749A1 (en) * 2014-08-18 2016-02-18 Dr. Ing. H.C. F. Porsche Aktiengesellschaft Method and device for controlling a technical device
US9778749B2 (en) 2014-08-22 2017-10-03 Google Inc. Occluded gesture recognition
US11169988B2 (en) 2014-08-22 2021-11-09 Google Llc Radar recognition-aided search
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
CN106797493A (en) 2014-09-02 2017-05-31 苹果公司 Music user interface
US10642365B2 (en) 2014-09-09 2020-05-05 Microsoft Technology Licensing, Llc Parametric inertia and APIs
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US9329715B2 (en) 2014-09-11 2016-05-03 Qeexo, Co. Method and apparatus for differentiating touch screen users based on touch event analysis
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US11619983B2 (en) 2014-09-15 2023-04-04 Qeexo, Co. Method and apparatus for resolving touch screen ambiguities
KR20160034776A (en) 2014-09-22 2016-03-30 삼성전자주식회사 Device and method of controlling the device
US10296206B2 (en) * 2014-09-23 2019-05-21 Microsoft Technology Licensing, Llc Multi-finger touchpad gestures
US9769227B2 (en) 2014-09-24 2017-09-19 Microsoft Technology Licensing, Llc Presentation of computing environment on multiple devices
US10635296B2 (en) * 2014-09-24 2020-04-28 Microsoft Technology Licensing, Llc Partitioned application presentation across devices
US10448111B2 (en) 2014-09-24 2019-10-15 Microsoft Technology Licensing, Llc Content projection
US10606417B2 (en) 2014-09-24 2020-03-31 Qeexo, Co. Method for improving accuracy of touch screen event analysis by use of spatiotemporal touch patterns
US10025684B2 (en) 2014-09-24 2018-07-17 Microsoft Technology Licensing, Llc Lending target device resources to host device computing environment
US10282024B2 (en) 2014-09-25 2019-05-07 Qeexo, Co. Classifying contacts or associations with a touch sensitive device
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
KR101651006B1 (en) * 2014-09-30 2016-08-24 원혁 Touch screen apparatus and method for operating touch screen apparatus
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9600080B2 (en) 2014-10-02 2017-03-21 Google Inc. Non-line-of-sight radar-based gesture recognition
WO2016065568A1 (en) 2014-10-30 2016-05-06 Microsoft Technology Licensing, Llc Multi-configuration input device
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
JP6055459B2 (en) * 2014-12-17 2016-12-27 京セラドキュメントソリューションズ株式会社 Touch panel device and image processing device
US20180267637A1 (en) * 2014-12-22 2018-09-20 Volkswagen Ag Finger-operated control bar, and use of the finger-operated control bar
US20160179355A1 (en) * 2014-12-23 2016-06-23 General Electric Company System and method for managing image scan parameters in medical imaging
US11182023B2 (en) 2015-01-28 2021-11-23 Flatfrog Laboratories Ab Dynamic touch quarantine frames
US10318074B2 (en) 2015-01-30 2019-06-11 Flatfrog Laboratories Ab Touch-sensing OLED display with tilted emitters
WO2016130074A1 (en) 2015-02-09 2016-08-18 Flatfrog Laboratories Ab Optical touch system comprising means for projecting and detecting light beams above and inside a transmissive panel
US10401546B2 (en) 2015-03-02 2019-09-03 Flatfrog Laboratories Ab Optical component for light coupling
KR102264748B1 (en) * 2015-03-03 2021-06-14 삼성디스플레이 주식회사 Touch device and display including the same
US20160259544A1 (en) * 2015-03-04 2016-09-08 Artem Polikarpov Systems And Methods For Virtual Periphery Interaction
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9645732B2 (en) 2015-03-08 2017-05-09 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US10048757B2 (en) 2015-03-08 2018-08-14 Apple Inc. Devices and methods for controlling media presentation
US9990107B2 (en) 2015-03-08 2018-06-05 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9632664B2 (en) 2015-03-08 2017-04-25 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10095396B2 (en) 2015-03-08 2018-10-09 Apple Inc. Devices, methods, and graphical user interfaces for interacting with a control object while dragging another object
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9639184B2 (en) 2015-03-19 2017-05-02 Apple Inc. Touch input cursor manipulation
US9785305B2 (en) 2015-03-19 2017-10-10 Apple Inc. Touch input cursor manipulation
US20170045981A1 (en) 2015-08-10 2017-02-16 Apple Inc. Devices and Methods for Processing Touch Inputs Based on Their Intensities
US10152208B2 (en) 2015-04-01 2018-12-11 Apple Inc. Devices and methods for processing touch inputs based on their intensities
US9727151B2 (en) 2015-04-16 2017-08-08 Google Inc. Avoiding accidental cursor movement when contacting a surface of a trackpad
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
EP3289432B1 (en) 2015-04-30 2019-06-12 Google LLC Rf-based micro-motion tracking for gesture tracking and recognition
TWI588734B (en) * 2015-05-26 2017-06-21 仁寶電腦工業股份有限公司 Electronic apparatus and method for operating electronic apparatus
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10088908B1 (en) 2015-05-27 2018-10-02 Google Llc Gesture detection and interactions
JP6314914B2 (en) * 2015-06-04 2018-04-25 京セラドキュメントソリューションズ株式会社 Image forming apparatus and operation screen control method of image forming apparatus
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US9891811B2 (en) 2015-06-07 2018-02-13 Apple Inc. Devices and methods for navigating between user interfaces
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10200598B2 (en) 2015-06-07 2019-02-05 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9674426B2 (en) 2015-06-07 2017-06-06 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US10346030B2 (en) 2015-06-07 2019-07-09 Apple Inc. Devices and methods for navigating between user interfaces
US9860451B2 (en) 2015-06-07 2018-01-02 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US10795558B2 (en) 2015-06-07 2020-10-06 Apple Inc. Device, method, and graphical user interface for providing and interacting with a virtual drawing aid
US9830048B2 (en) 2015-06-07 2017-11-28 Apple Inc. Devices and methods for processing touch inputs with instructions in a web page
US9961239B2 (en) 2015-06-07 2018-05-01 Apple Inc. Touch accommodation options
US10466119B2 (en) 2015-06-10 2019-11-05 Nextinput, Inc. Ruggedized wafer level MEMS force sensor with a tolerance trench
US10599903B2 (en) 2015-06-26 2020-03-24 Lenovo (Beijing) Limited Information processing method and electronic device
WO2017019028A1 (en) * 2015-07-28 2017-02-02 Hewlett Packard Enterprise Development Lp Application launch state determination
US9965445B2 (en) * 2015-08-06 2018-05-08 FiftyThree, Inc. Systems and methods for gesture-based formatting
US10416800B2 (en) 2015-08-10 2019-09-17 Apple Inc. Devices, methods, and graphical user interfaces for adjusting user interface objects
US10235035B2 (en) 2015-08-10 2019-03-19 Apple Inc. Devices, methods, and graphical user interfaces for content navigation and manipulation
US9880735B2 (en) 2015-08-10 2018-01-30 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10248308B2 (en) 2015-08-10 2019-04-02 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interfaces with physical gestures
CN105117100A (en) * 2015-08-19 2015-12-02 小米科技有限责任公司 Target object display method and apparatus
US10642404B2 (en) 2015-08-24 2020-05-05 Qeexo, Co. Touch sensitive device with multi-sensor stream synchronized data
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
CN105162982A (en) * 2015-09-17 2015-12-16 小米科技有限责任公司 Answering extension function display method and device
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US11150774B2 (en) * 2015-09-25 2021-10-19 International Business Machines Corporation Modifying display of objects on a user interface for a computing device based on detected patterns of user interaction
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10817065B1 (en) 2015-10-06 2020-10-27 Google Llc Gesture recognition using multiple antenna
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
EP3387516B1 (en) 2015-12-09 2022-04-20 FlatFrog Laboratories AB Improved stylus identification
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10528186B2 (en) * 2016-03-31 2020-01-07 Rovi Guides, Inc. Systems and methods for controlling playback of a media asset using a touch screen
US10492302B2 (en) 2016-05-03 2019-11-26 Google Llc Connecting an electronic component to an interactive textile
WO2017200949A1 (en) 2016-05-16 2017-11-23 Google Llc Interactive fabric
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US10303255B2 (en) * 2016-05-31 2019-05-28 Atheer, Inc. Method and apparatus for unified free space pose input system navigation
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179588B1 (en) 2016-06-09 2019-02-22 Apple Inc. Intelligent automated assistant in a home environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10637986B2 (en) 2016-06-10 2020-04-28 Apple Inc. Displaying and updating a set of application views
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
AU2017100670C4 (en) 2016-06-12 2019-11-21 Apple Inc. User interfaces for retrieving contextually relevant media content
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10817126B2 (en) * 2016-09-20 2020-10-27 Apple Inc. 3D document editing system
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
EP3545392A4 (en) 2016-11-24 2020-07-29 FlatFrog Laboratories AB Automatic optimisation of touch signal
US10579150B2 (en) * 2016-12-05 2020-03-03 Google Llc Concurrent detection of absolute distance and relative movement for sensing action gestures
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
EP4152132A1 (en) 2016-12-07 2023-03-22 FlatFrog Laboratories AB An improved touch device
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
WO2018141948A1 (en) 2017-02-06 2018-08-09 Flatfrog Laboratories Ab Optical coupling in touch-sensing systems
EP3580539A4 (en) 2017-02-09 2020-11-25 Nextinput, Inc. Integrated digital force sensors and related methods of manufacture
US11243125B2 (en) 2017-02-09 2022-02-08 Nextinput, Inc. Integrated piezoresistive and piezoelectric fusion force sensor
CN106843737B (en) * 2017-02-13 2020-05-08 北京新美互通科技有限公司 Text input method and device and terminal equipment
US10684758B2 (en) 2017-02-20 2020-06-16 Microsoft Technology Licensing, Llc Unified system for bimanual interactions
US10558341B2 (en) * 2017-02-20 2020-02-11 Microsoft Technology Licensing, Llc Unified system for bimanual interactions on flexible representations of content
US10572545B2 (en) 2017-03-03 2020-02-25 Perkinelmer Informatics, Inc Systems and methods for searching and indexing documents comprising chemical information
WO2018174787A1 (en) 2017-03-22 2018-09-27 Flatfrog Laboratories Eraser for touch displays
EP3602259A4 (en) 2017-03-28 2021-01-20 FlatFrog Laboratories AB Touch sensing apparatus and method for assembly
US10992795B2 (en) 2017-05-16 2021-04-27 Apple Inc. Methods and interfaces for home media control
US11431836B2 (en) 2017-05-02 2022-08-30 Apple Inc. Methods and interfaces for initiating media playback
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK201770427A1 (en) 2017-05-12 2018-12-20 Apple Inc. Low-latency intelligent automated assistant
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
US10928980B2 (en) 2017-05-12 2021-02-23 Apple Inc. User interfaces for playing and managing audio items
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services
CN111343060B (en) 2017-05-16 2022-02-11 苹果公司 Method and interface for home media control
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US20220279063A1 (en) 2017-05-16 2022-09-01 Apple Inc. Methods and interfaces for home media control
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US11221263B2 (en) 2017-07-19 2022-01-11 Nextinput, Inc. Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die
US11423686B2 (en) 2017-07-25 2022-08-23 Qorvo Us, Inc. Integrated fingerprint and force sensor
WO2019023552A1 (en) 2017-07-27 2019-01-31 Nextinput, Inc. A wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture
US11301124B2 (en) 2017-08-18 2022-04-12 Microsoft Technology Licensing, Llc User interface modification using preview panel
US20190056857A1 (en) * 2017-08-18 2019-02-21 Microsoft Technology Licensing, Llc Resizing an active region of a user interface
US11237699B2 (en) 2017-08-18 2022-02-01 Microsoft Technology Licensing, Llc Proximal menu generation
US11256371B2 (en) 2017-09-01 2022-02-22 Flatfrog Laboratories Ab Optical component
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
CN107728918A (en) * 2017-09-27 2018-02-23 北京三快在线科技有限公司 Browse the method, apparatus and electronic equipment of continuous page
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
CN109634450A (en) * 2017-10-09 2019-04-16 华为技术有限公司 False-touch prevention detection method, device and terminal
WO2019079420A1 (en) 2017-10-17 2019-04-25 Nextinput, Inc. Temperature coefficient of offset compensation for force sensor and strain gauge
CN111295877B (en) * 2017-11-01 2022-01-11 富士胶片株式会社 Image pickup apparatus, method of operating the same, and computer readable medium
US11385108B2 (en) 2017-11-02 2022-07-12 Nextinput, Inc. Sealed force sensor with etch stop layer
US11874185B2 (en) 2017-11-16 2024-01-16 Nextinput, Inc. Force attenuator for force sensor
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
CN110069147B (en) * 2018-01-23 2023-02-03 可赛尔内存股份有限公司 Control device and control method thereof
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
WO2019172826A1 (en) 2018-03-05 2019-09-12 Flatfrog Laboratories Ab Improved touch-sensing apparatus
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US11669243B2 (en) 2018-06-03 2023-06-06 Apple Inc. Systems and methods for activating and using a trackpad at an electronic device with a touch-sensitive display and no force sensors
US11076039B2 (en) 2018-06-03 2021-07-27 Apple Inc. Accelerated task performance
US10776006B2 (en) 2018-06-03 2020-09-15 Apple Inc. Systems and methods for activating and using a trackpad at an electronic device with a touch-sensitive display and no force sensors
US11009989B2 (en) 2018-08-21 2021-05-18 Qeexo, Co. Recognizing and rejecting unintentional touch events associated with a touch sensitive device
US11204682B1 (en) * 2018-10-29 2021-12-21 State Farm Mutual Automobile Insurance Company Dynamic data-driven consolidation of user interface interactions requesting roadside assistance
WO2020100081A1 (en) * 2018-11-14 2020-05-22 Wix.Com Ltd. System and method for creation and handling of configurable applications for website building systems
US10962427B2 (en) 2019-01-10 2021-03-30 Nextinput, Inc. Slotted MEMS force sensor
US11943563B2 (en) 2019-01-25 2024-03-26 FlatFrog Laboratories, AB Videoconferencing terminal and method of operating the same
US10942603B2 (en) 2019-05-06 2021-03-09 Qeexo, Co. Managing activity states of an application processor in relation to touch or hover interactions with a touch sensitive device
DK201970535A1 (en) 2019-05-06 2020-12-21 Apple Inc Media browsing user interface with intelligently selected representative media items
DK201970533A1 (en) 2019-05-31 2021-02-15 Apple Inc Methods and user interfaces for sharing audio
AU2020282362B2 (en) 2019-05-31 2022-01-13 Apple Inc. User interfaces for audio media control
US11010121B2 (en) 2019-05-31 2021-05-18 Apple Inc. User interfaces for audio media control
US11379113B2 (en) 2019-06-01 2022-07-05 Apple Inc. Techniques for selecting text
US11194467B2 (en) 2019-06-01 2021-12-07 Apple Inc. Keyboard management user interfaces
US11980792B2 (en) 2019-06-05 2024-05-14 Qeexo, Co. Method and apparatus for calibrating a user activity model used by a mobile device
US20220197487A1 (en) * 2019-06-19 2022-06-23 Ja Beom KOO View mode change device using touch pattern input and method therefor
US11231815B2 (en) 2019-06-28 2022-01-25 Qeexo, Co. Detecting object proximity using touch sensitive surface sensing and ultrasonic sensing
US11592423B2 (en) 2020-01-29 2023-02-28 Qeexo, Co. Adaptive ultrasonic sensing techniques and systems to mitigate interference
US11893189B2 (en) 2020-02-10 2024-02-06 Flatfrog Laboratories Ab Touch-sensing apparatus
DK202070616A1 (en) 2020-02-14 2022-01-14 Apple Inc User interfaces for workout content
US11513667B2 (en) 2020-05-11 2022-11-29 Apple Inc. User interface for audio message
EP4154096A1 (en) 2020-05-18 2023-03-29 Apple Inc. User interfaces for viewing and refining the current location of an electronic device
US11416136B2 (en) 2020-09-14 2022-08-16 Apple Inc. User interfaces for assigning and responding to user inputs
US11392291B2 (en) 2020-09-25 2022-07-19 Apple Inc. Methods and interfaces for media control with dynamic feedback

Family Cites Families (716)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3333160A (en) 1964-02-24 1967-07-25 Water Economy And Res Company Proximity responsive system
US3541541A (en) 1967-06-21 1970-11-17 Stanford Research Inst X-y position indicator for a display system
US3609695A (en) 1968-05-10 1971-09-28 Honeywell Inc Display-entry data terminal
US3662105A (en) 1970-05-21 1972-05-09 Univ Kentucky Res Found Electrical sensor of plane coordinates
US3757322A (en) 1971-02-03 1973-09-04 Hall Barkan Instr Inc Transparent touch controlled interface with interreactively related display
US3846826A (en) 1971-08-12 1974-11-05 R Mueller Direct television drawing and image manipulating system
US3798370A (en) 1972-04-17 1974-03-19 Elographics Inc Electrographic sensor for determining planar coordinates
US3748751A (en) 1972-09-07 1973-07-31 Us Navy Laser machine gun simulator
US3825730A (en) 1972-11-07 1974-07-23 Searle Medidata Inc Two-way projection system for data input
JPS51112236A (en) 1975-03-28 1976-10-04 Hitachi Ltd Shape position recognizer unit
US4017848A (en) 1975-05-19 1977-04-12 Rockwell International Corporation Transparent keyboard switch and array
US4146924A (en) 1975-09-22 1979-03-27 Board Of Regents For Education Of The State Of Rhode Island System for visually determining position in space and/or orientation in space and apparatus employing same
FR2392548A2 (en) 1977-05-25 1978-12-22 Serras Paulet Edouard PUSH-BUTTON DEVICE FOR SWITCHING AN ELECTRONIC OR ELECTRIC CIRCUIT
US4202041A (en) 1977-07-11 1980-05-06 Massachusetts Institute Of Technology Dynamically variable keyboard terminal
CA1103803A (en) 1978-03-01 1981-06-23 National Research Council Of Canada Method and apparatus of determining the center of area or centroid of a geometrical area of unspecified shape lying in a larger x-y scan field
US4246452A (en) 1979-01-05 1981-01-20 Mattel, Inc. Switch apparatus
US4305131A (en) 1979-02-05 1981-12-08 Best Robert M Dialog between TV movies and human viewers
JPS56132505A (en) 1980-03-24 1981-10-16 Hitachi Ltd Position detecting method
US4305071A (en) 1980-04-16 1981-12-08 Bell Telephone Laboratories, Incorporated Touch sensitive screen signal detection arrangement
US4346376A (en) 1980-04-16 1982-08-24 Bell Telephone Laboratories, Incorporated Touch position sensitive surface
US4484179A (en) 1980-04-16 1984-11-20 At&T Bell Laboratories Touch position sensitive surface
US4686374A (en) 1980-06-26 1987-08-11 Diffracto Ltd. Surface reflectivity detector with oil mist reflectivity enhancement
US4375674A (en) 1980-10-17 1983-03-01 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Kinesimetric method and apparatus
US4396945A (en) 1981-08-19 1983-08-02 Solid Photography Inc. Method of sensing the position and orientation of elements in space
US4475122A (en) 1981-11-09 1984-10-02 Tre Semiconductor Equipment Corporation Automatic wafer alignment technique
US4542375A (en) 1982-02-11 1985-09-17 At&T Bell Laboratories Deformable touch sensitive surface
US4613942A (en) 1982-02-19 1986-09-23 Chen Richard M Orientation and control system for robots
US4549279A (en) 1983-01-21 1985-10-22 The Laitram Corporation Single hand, single finger stroke alphameric data processing keyboard system
US4631525A (en) 1983-04-11 1986-12-23 Sony Corporation Digital fader or like device
US4631676A (en) 1983-05-25 1986-12-23 Hospital For Joint Diseases Or Computerized video gait and motion analysis system and method
US4644326A (en) 1983-06-03 1987-02-17 Secure Keyboards Limited Unitary key panel
US4654872A (en) 1983-07-25 1987-03-31 Omron Tateisi Electronics Co. System for recognizing three-dimensional objects
JPS6029833A (en) 1983-07-28 1985-02-15 Canon Inc Image display device
US4561017A (en) 1983-08-19 1985-12-24 Richard Greene Graphic input apparatus
US4550221A (en) 1983-10-07 1985-10-29 Scott Mabusth Touch sensitive control device
US4733222A (en) 1983-12-27 1988-03-22 Integrated Touch Arrays, Inc. Capacitance-variation-sensitive touch sensing array system
US4629319A (en) 1984-02-14 1986-12-16 Diffracto Ltd. Panel surface flaw inspection
US4672364A (en) 1984-06-18 1987-06-09 Carroll Touch Inc Touch input device having power profiling
US4672558A (en) 1984-09-25 1987-06-09 Aquila Technologies Group, Inc. Touch-sensitive data input device
US4692809A (en) 1984-11-20 1987-09-08 Hughes Aircraft Company Integrated touch paint system for displays
US4695827A (en) 1984-11-20 1987-09-22 Hughes Aircraft Company Electromagnetic energy interference seal for light beam touch panels
US4853888A (en) 1984-12-31 1989-08-01 The Boeing Company Programmable multifunction keyboard
US4644100A (en) 1985-03-22 1987-02-17 Zenith Electronics Corporation Surface acoustic wave touch panel system
US4710760A (en) 1985-03-07 1987-12-01 American Telephone And Telegraph Company, At&T Information Systems Inc. Photoelastic touch-sensitive screen
US4771276A (en) 1985-04-15 1988-09-13 International Business Machines Corporation Electromagnetic touch sensor input system in a cathode ray tube display device
US5017030A (en) 1986-07-07 1991-05-21 Crews Jay A Ergonomically designed keyboard
US4703306A (en) 1986-09-26 1987-10-27 The Maytag Company Appliance system
US4763356A (en) 1986-12-11 1988-08-09 AT&T Information Systems, Inc. American Telephone and Telegraph Company Touch screen form entry system
FR2608753B1 (en) 1986-12-18 1991-06-07 Centre Nat Rech Scient DEVICE FOR TWO-DIMENSIONAL LOCATION OF CURRENT GENERATING EVENTS ON A RESISTIVE SURFACE
US4787040A (en) 1986-12-22 1988-11-22 International Business Machines Corporation Display system for automotive vehicle
JPS63167923A (en) 1987-01-05 1988-07-12 Pfu Ltd Image data input device
US4746770A (en) 1987-02-17 1988-05-24 Sensor Frame Incorporated Method and apparatus for isolating and manipulating graphic objects on computer video monitor
FR2612326A1 (en) 1987-03-13 1988-09-16 Thomson Csf METHOD OF ADJUSTING THE COLORS OF A POLYCHROME PLASMA PANEL AND PLASMA PANEL USING THE SAME
US4988981B1 (en) 1987-03-17 1999-05-18 Vpl Newco Inc Computer data entry and manipulation apparatus and method
GB2204131B (en) 1987-04-28 1991-04-17 Ibm Graphics input tablet
US4806846A (en) 1987-07-06 1989-02-21 Kerber George L High accuracy direct reading capacitance-to-voltage converter
US4814759A (en) 1987-07-08 1989-03-21 Clinicom Incorporated Flat panel display monitor apparatus
US4772028A (en) 1987-08-27 1988-09-20 Rockhold Christopher K Electronic shootout game
US5153829A (en) 1987-11-11 1992-10-06 Canon Kabushiki Kaisha Multifunction musical information processing apparatus
US4914624A (en) 1988-05-06 1990-04-03 Dunthorn David I Virtual button for touch screen
CH672704B5 (en) 1988-05-26 1990-06-29 Ebauchesfabrik Eta Ag
US5060135A (en) 1988-09-16 1991-10-22 Wang Laboratories, Inc. Apparatus for manipulating documents in a data processing system utilizing reduced images of sheets of information which are movable
US4968877A (en) 1988-09-14 1990-11-06 Sensor Frame Corporation VideoHarp
CA2002912A1 (en) 1988-11-14 1990-05-14 William A. Clough Portable computer with touch screen and computer system employing same
US5045843B1 (en) 1988-12-06 1996-07-16 Selectech Ltd Optical pointing device
US5741990A (en) 1989-02-17 1998-04-21 Notepool, Ltd. Method of and means for producing musical note relationships
US4898555A (en) 1989-03-23 1990-02-06 Bell & Howell Publication Systems Company Display screen bezel and assembly method
JPH0649030Y2 (en) 1989-04-05 1994-12-12 パイオニア株式会社 Data input device
US5620267A (en) 1993-10-15 1997-04-15 Keyboard Advancements, Inc. Keyboard with thumb activated control key
US4948371A (en) 1989-04-25 1990-08-14 The United States Of America As Represented By The United States Department Of Energy System for training and evaluation of security personnel in use of firearms
US5252951A (en) * 1989-04-28 1993-10-12 International Business Machines Corporation Graphical user interface with gesture recognition in a multiapplication environment
US5157384A (en) 1989-04-28 1992-10-20 International Business Machines Corporation Advanced user interface
DK174788B1 (en) 1989-05-08 2003-11-10 Dainippon Printing Co Ltd A rear projection screen
US5072294A (en) 1989-06-07 1991-12-10 Loredan Biomedical, Inc. Method and apparatus for analyzing a body having a marker located thereon
US5305017A (en) 1989-08-16 1994-04-19 Gerpheide George E Methods and apparatus for data input
US5189403A (en) 1989-09-26 1993-02-23 Home Row, Inc. Integrated keyboard and pointing device system with automatic mode change
CA2027103A1 (en) 1989-10-13 1991-04-14 William A. Clough Method and apparatus for displaying simulated keyboards on touch-sensitive displays
JP2622620B2 (en) 1989-11-07 1997-06-18 プロクシマ コーポレイション Computer input system for altering a computer generated display visible image
JP2854359B2 (en) 1990-01-24 1999-02-03 富士通株式会社 Image processing system
US5241308A (en) 1990-02-22 1993-08-31 Paragon Systems, Inc. Force sensitive touch panel
US5581243A (en) 1990-06-04 1996-12-03 Microslate Inc. Method and apparatus for displaying simulated keyboards on touch-sensitive displays
US5097425A (en) 1990-06-11 1992-03-17 Semantic Compaction Systems Predictive scanning input system for rapid selection of visual indicators
JP3301079B2 (en) 1990-06-18 2002-07-15 ソニー株式会社 Information input device, information input method, information processing device, and information processing method
JPH0454523A (en) 1990-06-21 1992-02-21 Hitachi Aic Inc Touch panel
GB2245708A (en) 1990-06-29 1992-01-08 Philips Electronic Associated Touch sensor array systems
CA2087523C (en) 1990-07-17 1997-04-15 Mark Andrew Shackleton Method of processing an image
GB9019538D0 (en) 1990-09-07 1990-10-24 Philips Electronic Associated Tracking a moving object
US5224861A (en) 1990-09-17 1993-07-06 Hughes Aircraft Company Training device onboard instruction station
US5119079A (en) 1990-09-17 1992-06-02 Xerox Corporation Touch screen user interface with expanding touch locations for a reprographic machine
US5347295A (en) 1990-10-31 1994-09-13 Go Corporation Control of a computer through a position-sensed stylus
US5148155A (en) 1990-11-13 1992-09-15 Wang Laboratories, Inc. Computer with tablet input to standard programs
US5260697A (en) 1990-11-13 1993-11-09 Wang Laboratories, Inc. Computer with separate display plane and user interface processor
US5442742A (en) 1990-12-21 1995-08-15 Apple Computer, Inc. Method and apparatus for the manipulation of text on a computer display screen
KR100318330B1 (en) 1991-04-08 2002-04-22 가나이 쓰도무 Monitoring device
US5689253A (en) 1991-04-10 1997-11-18 Kinesis Corporation Ergonomic keyboard apparatus
US5898434A (en) 1991-05-15 1999-04-27 Apple Computer, Inc. User interface system having programmable user interface elements
US5548667A (en) 1991-05-24 1996-08-20 Sony Corporation Image processing system and method thereof in which three dimensional shape is reproduced from two dimensional image data
US5178477A (en) 1991-06-06 1993-01-12 Gambaro Thomas L Ergonomic keyboard input device
US5168531A (en) 1991-06-27 1992-12-01 Digital Equipment Corporation Real-time recognition of pointing information from video
WO1993000629A1 (en) 1991-06-28 1993-01-07 Infogrip, Inc. Chordic keyboard system
US5666113A (en) 1991-07-31 1997-09-09 Microtouch Systems, Inc. System for using a touchpad input device for cursor control and keyboard emulation
US5227985A (en) 1991-08-19 1993-07-13 University Of Maryland Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monitored object
JP2827612B2 (en) 1991-10-07 1998-11-25 富士通株式会社 A touch panel device and a method for displaying an object on the touch panel device.
US5335557A (en) 1991-11-26 1994-08-09 Taizo Yasutake Touch sensitive input control device
US5963671A (en) 1991-11-27 1999-10-05 International Business Machines Corporation Enhancement of soft keyboard operations using trigram prediction
CA2078607A1 (en) 1991-12-13 1993-06-14 Thomas H. Speeter Intelligent work surfaces
US5212555A (en) 1991-12-17 1993-05-18 Texas Instruments Incorporated Image capture with spatial light modulator and single-cell photosensor
JPH0619663A (en) 1991-12-27 1994-01-28 Nippon Telegr & Teleph Corp <Ntt> Automatic control method for multiwindow
US5281966A (en) 1992-01-31 1994-01-25 Walsh A Peter Method of encoding alphabetic characters for a chord keyboard
EP0554492B1 (en) 1992-02-07 1995-08-09 International Business Machines Corporation Method and device for optical input of commands or data
US5483261A (en) * 1992-02-14 1996-01-09 Itu Research, Inc. Graphical input controller and method with rear screen image detection
EP0563477A1 (en) 1992-03-25 1993-10-06 Visage Inc. Touch screen sensing apparatus
US5495269A (en) 1992-04-03 1996-02-27 Xerox Corporation Large area electronic writing system
US5347629A (en) 1992-04-07 1994-09-13 International Business Machines Corporation Graphical user interface including updating of multiple panels using what you see is what you get (WYSIWYG) editor
US5398310A (en) 1992-04-13 1995-03-14 Apple Computer, Incorporated Pointing gesture based computer note pad paging and scrolling interface
US5563996A (en) 1992-04-13 1996-10-08 Apple Computer, Inc. Computer note pad including gesture based note division tools and method
JPH05297979A (en) 1992-04-20 1993-11-12 Matsushita Electric Ind Co Ltd Laptop computer
US5523775A (en) 1992-05-26 1996-06-04 Apple Computer, Inc. Method for selecting objects on a computer display
US5592566A (en) 1992-05-27 1997-01-07 Apple Computer, Incorporated Method and apparatus for computerized recognition
US5544295A (en) 1992-05-27 1996-08-06 Apple Computer, Inc. Method and apparatus for indicating a change in status of an object and its disposition using animation
US5710844A (en) 1992-05-27 1998-01-20 Apple Computer Method for searching and displaying results in a pen-based computer system
US5452371A (en) 1992-05-27 1995-09-19 Apple Computer, Inc. Method of aligning shapes on a display of a computer system
US5463696A (en) 1992-05-27 1995-10-31 Apple Computer, Inc. Recognition system and method for user inputs to a computer system
US5764818A (en) 1992-05-27 1998-06-09 Apple Computer, Inc. Method for locating and displaying information in a pointer-based computer system
US5543590A (en) 1992-06-08 1996-08-06 Synaptics, Incorporated Object position detector with edge motion feature
US5543591A (en) 1992-06-08 1996-08-06 Synaptics, Incorporated Object position detector with edge motion feature and gesture recognition
US5488204A (en) 1992-06-08 1996-01-30 Synaptics, Incorporated Paintbrush stylus for capacitive touch sensor pad
EP0574213B1 (en) 1992-06-08 1999-03-24 Synaptics, Inc. Object position detector
US6028271A (en) 1992-06-08 2000-02-22 Synaptics, Inc. Object position detector with edge motion feature and gesture recognition
US5880411A (en) 1992-06-08 1999-03-09 Synaptics, Incorporated Object position detector with edge motion feature and gesture recognition
US5565658A (en) 1992-07-13 1996-10-15 Cirque Corporation Capacitance-based proximity with interference rejection apparatus and methods
US5328190A (en) 1992-08-04 1994-07-12 Dart International, Inc. Method and apparatus enabling archery practice
US5319386A (en) 1992-08-04 1994-06-07 Gunn Gary J Ideographic character selection method and apparatus
US5418760A (en) 1992-08-18 1995-05-23 Casio Computer Co., Ltd. Electronic devices with a liquid crystal display
US5821930A (en) 1992-08-23 1998-10-13 U S West, Inc. Method and system for generating a working window in a computer system
JP3268467B2 (en) 1992-09-08 2002-03-25 株式会社日立製作所 Telephone
US7098891B1 (en) 1992-09-18 2006-08-29 Pryor Timothy R Method for providing human input to a computer
US6008800A (en) 1992-09-18 1999-12-28 Pryor; Timothy R. Man machine interfaces for entering data into a computer
US5982352A (en) 1992-09-18 1999-11-09 Pryor; Timothy R. Method for providing human input to a computer
US7084859B1 (en) 1992-09-18 2006-08-01 Pryor Timothy R Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
US5345543A (en) 1992-11-16 1994-09-06 Apple Computer, Inc. Method for manipulating objects on a computer display
JP3469912B2 (en) 1992-11-18 2003-11-25 株式会社デジタル Touch panel input device and input method capable of multiple simultaneous inputs
US5363051A (en) 1992-11-23 1994-11-08 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Steering capaciflector sensor
US5612719A (en) 1992-12-03 1997-03-18 Apple Computer, Inc. Gesture sensitive buttons for graphical user interfaces
US5412189A (en) 1992-12-21 1995-05-02 International Business Machines Corporation Touch screen apparatus with tactile information
US5568604A (en) 1992-12-31 1996-10-22 U S West Technologies, Inc. Method and system for generating a working window in a computer system
US5463725A (en) 1992-12-31 1995-10-31 International Business Machines Corp. Data processing system graphical user interface which emulates printed material
US5513309A (en) 1993-01-05 1996-04-30 Apple Computer, Inc. Graphic editor user interface for a pointer-based computer system
US5495576A (en) 1993-01-11 1996-02-27 Ritchey; Kurtis J. Panoramic image based virtual reality/telepresence audio-visual system and method
US5463388A (en) 1993-01-29 1995-10-31 At&T Ipm Corp. Computer mouse or keyboard input device utilizing capacitive sensors
JP2986047B2 (en) 1993-04-29 1999-12-06 インターナショナル・ビジネス・マシーンズ・コーポレイション Digital input display device and input processing device and method
US5563632A (en) 1993-04-30 1996-10-08 Microtouch Systems, Inc. Method of and apparatus for the elimination of the effects of internal interference in force measurement systems, including touch - input computer and related displays employing touch force location measurement techniques
EP0622722B1 (en) 1993-04-30 2002-07-17 Xerox Corporation Interactive copying system
US5677710A (en) 1993-05-10 1997-10-14 Apple Computer, Inc. Recognition keypad
US5808567A (en) 1993-05-17 1998-09-15 Dsi Datotech Systems, Inc. Apparatus and method of communicating using three digits of a hand
DE69432199T2 (en) 1993-05-24 2004-01-08 Sun Microsystems, Inc., Mountain View Graphical user interface with methods for interfacing with remote control devices
WO1994029788A1 (en) 1993-06-15 1994-12-22 Honeywell Inc. A method for utilizing a low resolution touch screen system in a high resolution graphics environment
US5367453A (en) 1993-08-02 1994-11-22 Apple Computer, Inc. Method and apparatus for correcting words
GB9317932D0 (en) 1993-08-26 1993-10-13 Cambridge Display Tech Ltd Electroluminescent devices
JPH07230352A (en) 1993-09-16 1995-08-29 Hitachi Ltd Touch position detecting device and touch instruction processor
US5583946A (en) 1993-09-30 1996-12-10 Apple Computer, Inc. Method and apparatus for recognizing gestures on a computer system
US5510806A (en) 1993-10-28 1996-04-23 Dell Usa, L.P. Portable computer having an LCD projection display system
US5422656A (en) 1993-11-01 1995-06-06 International Business Machines Corp. Personal communicator having improved contrast control for a liquid crystal, touch sensitive display
US6243071B1 (en) 1993-11-03 2001-06-05 Apple Computer, Inc. Tool set for navigating through an electronic book
JPH07129312A (en) 1993-11-05 1995-05-19 Oki Electric Ind Co Ltd Picture processor
US5982355A (en) 1993-11-05 1999-11-09 Jaeger; Denny Multiple purpose controls for electrical systems
US5841428A (en) 1993-11-05 1998-11-24 Intertactile Technologies Corporation Rotary circuit control devices with changeable graphics
US5805145A (en) 1993-11-05 1998-09-08 Intertactile Technologies Corporation Circuit control panel displaying changeable graphics
WO1995012877A1 (en) 1993-11-05 1995-05-11 Denny Jaeger Operator/circuit interface with integrated display screen
US5936613A (en) 1993-11-05 1999-08-10 Intertactile Technologies Corporation Rotary circuit control devices with changeable graphics
US5805146A (en) 1993-11-05 1998-09-08 Intertactile Technologies Corporation Integrated display screen and slidable control for electrical circuits
US5777603A (en) 1993-11-05 1998-07-07 Intertactile Technologies Corporation Flat panel display with optical signal transparent zone
US5530456A (en) 1993-12-09 1996-06-25 Nec Corporation Position information input method and device
JP3546337B2 (en) 1993-12-21 2004-07-28 ゼロックス コーポレイション User interface device for computing system and method of using graphic keyboard
DE69426919T2 (en) 1993-12-30 2001-06-28 Xerox Corp Apparatus and method for performing many chaining command gestures in a gesture user interface system
DE69428675T2 (en) 1993-12-30 2002-05-08 Xerox Corp Apparatus and method for supporting an implicit structuring of free-form lists, overviews, texts, tables and diagrams in an input system and editing system based on hand signals
US5471578A (en) 1993-12-30 1995-11-28 Xerox Corporation Apparatus and method for altering enclosure selections in a gesture based input system
US5572573A (en) 1994-01-25 1996-11-05 U S West Advanced Technologies, Inc. Removable user interface for use with interactive electronic devices
WO1995020787A1 (en) 1994-01-27 1995-08-03 Exos, Inc. Multimode feedback display technology
US5982302A (en) 1994-03-07 1999-11-09 Ure; Michael J. Touch-sensitive keyboard/mouse
GB9406702D0 (en) 1994-04-05 1994-05-25 Binstead Ronald P Multiple input proximity detector and touchpad system
JPH07275511A (en) 1994-04-06 1995-10-24 Sega Enterp Ltd Attraction development method for shooting game system
US5579036A (en) 1994-04-28 1996-11-26 At&T Global Information Solutions Company Touch screen device and shielding bracket therefor
WO1995031791A1 (en) 1994-05-12 1995-11-23 Apple Computer, Inc. Method and apparatus for noise filtering for an input device
GB2290896B (en) * 1994-06-13 1998-09-23 Nec Corp MOS four-quadrant multiplier
US5570113A (en) 1994-06-29 1996-10-29 International Business Machines Corporation Computer based pen system and method for automatically cancelling unwanted gestures and preventing anomalous signals as inputs to such system
DE4423005C1 (en) 1994-06-30 1995-11-30 Siemens Ag Computer data entry stylus with indistinguishable contact surfaces
US5530455A (en) 1994-08-10 1996-06-25 Mouse Systems Corporation Roller mouse for implementing scrolling in windows applications
US7489303B1 (en) 2001-02-22 2009-02-10 Pryor Timothy R Reconfigurable instrument panels
US9513744B2 (en) 1994-08-15 2016-12-06 Apple Inc. Control systems employing novel physical controls and touch screens
US5559301A (en) 1994-09-15 1996-09-24 Korg, Inc. Touchscreen interface having pop-up variable adjustment displays for controllers and audio processing systems
US5649706A (en) 1994-09-21 1997-07-22 Treat, Jr.; Erwin C. Simulator and practice method
WO1996009579A1 (en) 1994-09-22 1996-03-28 Izak Van Cruyningen Popup menus with directional gestures
EP0703525B1 (en) 1994-09-22 2001-12-05 Aisin Aw Co., Ltd. Touch display type information input system
US5572647A (en) 1994-11-04 1996-11-05 International Business Machines Corporation Visibility seeking scroll bars and other control constructs
US5745719A (en) 1995-01-19 1998-04-28 Falcon; Fernando D. Commands functions invoked from movement of a control input device
US5736976A (en) 1995-02-13 1998-04-07 Cheung; Nina T. Computer data entry apparatus with hand motion sensing and monitoring
US5565888A (en) 1995-02-17 1996-10-15 International Business Machines Corporation Method and apparatus for improving visibility and selectability of icons
US5594469A (en) 1995-02-21 1997-01-14 Mitsubishi Electric Information Technology Center America Inc. Hand gesture machine control system
US5838302A (en) 1995-02-24 1998-11-17 Casio Computer Co., Ltd. Data inputting devices for inputting typed and handwritten data in a mixed manner
US5748512A (en) 1995-02-28 1998-05-05 Microsoft Corporation Adjusting keyboard
US5618232A (en) 1995-03-23 1997-04-08 Martin; John R. Dual mode gaming device methods and systems
US5900876A (en) 1995-04-14 1999-05-04 Canon Kabushiki Kaisha Information processing apparatus and method with display book page turning
US5591945A (en) 1995-04-19 1997-01-07 Elo Touchsystems, Inc. Acoustic touch position sensor using higher order horizontally polarized shear wave propagation
US5729250A (en) 1995-05-08 1998-03-17 International Business Machines Corporation Front cover assembly for a touch sensitive device
US6255604B1 (en) 1995-05-31 2001-07-03 Canon Kabushiki Kaisha Coordinate detecting device for outputting coordinate data when two points are simultaneously depressed, method therefor and computer control device
US5689667A (en) 1995-06-06 1997-11-18 Silicon Graphics, Inc. Methods and system of controlling menus with radial and linear portions
US5798760A (en) 1995-06-07 1998-08-25 Vayda; Mark Radial graphical menuing system with concentric region menuing
US6031524A (en) 1995-06-07 2000-02-29 Intermec Ip Corp. Hand-held portable data terminal having removably interchangeable, washable, user-replaceable components with liquid-impervious seal
US5790107A (en) 1995-06-07 1998-08-04 Logitech, Inc. Touch sensing method and apparatus
US5502514A (en) 1995-06-07 1996-03-26 Nview Corporation Stylus position sensing and digital camera with a digital micromirror device
US5831601A (en) 1995-06-07 1998-11-03 Nview Corporation Stylus position sensing and digital camera with a digital micromirror device
US8228305B2 (en) 1995-06-29 2012-07-24 Apple Inc. Method for providing human input to a computer
US20090273574A1 (en) 1995-06-29 2009-11-05 Pryor Timothy R Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
JP3610635B2 (en) 1995-07-18 2005-01-19 マツダ株式会社 Display device for operation of in-vehicle equipment
JPH0935584A (en) 1995-07-21 1997-02-07 Yazaki Corp Display device for vehicle
US6005549A (en) 1995-07-24 1999-12-21 Forest; Donald K. User interface method and apparatus
US5999895A (en) 1995-07-24 1999-12-07 Forest; Donald K. Sound operated menu method and apparatus
US5702323A (en) 1995-07-26 1997-12-30 Poulton; Craig K. Electronic exercise enhancer
US5634102A (en) 1995-08-07 1997-05-27 Apple Computer, Inc. Methods and apparatus for a selectable backdrop
US5666502A (en) 1995-08-07 1997-09-09 Apple Computer, Inc. Graphical user interface using historical lists with field classes
US5745716A (en) 1995-08-07 1998-04-28 Apple Computer, Inc. Method and apparatus for tab access and tab cycling in a pen-based computer system
US5867149A (en) 1995-08-14 1999-02-02 Intertactile Technologies Corporation Switch key image display and operator/circuit interface
US5675361A (en) 1995-08-23 1997-10-07 Santilli; Donald S. Computer keyboard pointing device
US5746818A (en) 1995-08-31 1998-05-05 Seiko Epson Corporation Pigment ink composition capable of forming image having no significant bleeding or feathering
US6481851B1 (en) 1995-09-20 2002-11-19 Videotronic Systems Adjustable contrast reflected display system
US5636101A (en) 1995-09-27 1997-06-03 Dynapro Systems, Inc. Touch screen enclosure system having touch screen pan and hinged rear enclosure section for ease of serviceability
US5631805A (en) 1995-09-27 1997-05-20 Dynapro Systems, Inc. Touch screen enclosure having an insertable graphic sheet
US6049328A (en) 1995-10-20 2000-04-11 Wisconsin Alumni Research Foundation Flexible access system for touch screen devices
US5801941A (en) 1996-08-12 1998-09-01 International Business Machines Corporation Mobile client computer programmed to establish soft keyboard targeting sensitivity
JP3727399B2 (en) 1996-02-19 2005-12-14 ミサワホーム株式会社 Screen display type key input device
JPH09146708A (en) 1995-11-09 1997-06-06 Internatl Business Mach Corp <Ibm> Driving method for touch panel and touch input method
US6473069B1 (en) 1995-11-13 2002-10-29 Cirque Corporation Apparatus and method for tactile feedback from input device
US5767457A (en) 1995-11-13 1998-06-16 Cirque Corporation Apparatus and method for audible feedback from input device
US6107997A (en) 1996-06-27 2000-08-22 Ure; Michael J. Touch-sensitive keyboard/mouse and computing device using the same
EP0861485A1 (en) 1995-11-16 1998-09-02 Michael J. Ure Multi-touch input device, method and system that minimize the need for memorization
US5825308A (en) 1996-11-26 1998-10-20 Immersion Human Interface Corporation Force feedback interface having isotonic and isometric functionality
SE504758C2 (en) 1995-12-21 1997-04-21 Siemens Elema Ab Device Panel
US5730165A (en) 1995-12-26 1998-03-24 Philipp; Harald Time domain capacitive field detector
US5823782A (en) 1995-12-29 1998-10-20 Tinkers & Chance Character recognition educational system
US5920309A (en) 1996-01-04 1999-07-06 Logitech, Inc. Touch sensing method and apparatus
US5825352A (en) 1996-01-04 1998-10-20 Logitech, Inc. Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad
US5736975A (en) 1996-02-02 1998-04-07 Interactive Sales System Interactive video display
US5871251A (en) 1996-03-14 1999-02-16 Prince Corporation Visor display
WO1997040482A1 (en) 1996-04-24 1997-10-30 Logitech, Inc. Touch and pressure sensing method and apparatus
GB9608747D0 (en) 1996-04-26 1996-07-03 Philips Electronics Nv Fingerprint sensing devices and systems incorporating such
US5874948A (en) 1996-05-28 1999-02-23 International Business Machines Corporation Virtual pointing device for touchscreens
US5748184A (en) 1996-05-28 1998-05-05 International Business Machines Corporation Virtual pointing device for touchscreens
US5764222A (en) 1996-05-28 1998-06-09 International Business Machines Corporation Virtual pointing device for touchscreens
JP4484255B2 (en) 1996-06-11 2010-06-16 株式会社日立製作所 Information processing apparatus having touch panel and information processing method
US5956822A (en) 1996-06-12 1999-09-28 Knecht Filterwerke Gmbh Snap-lock connection for two parts which can rotate in relation to each other
US5835079A (en) 1996-06-13 1998-11-10 International Business Machines Corporation Virtual pointing device for touchscreens
US5808605A (en) 1996-06-13 1998-09-15 International Business Machines Corporation Virtual pointing device for touchscreens
US5856824A (en) 1996-06-25 1999-01-05 International Business Machines Corp. Reshapable pointing device for touchscreens
GB9613315D0 (en) 1996-06-25 1996-08-28 Philips Electronics Nv Virtual environment interaction and navigation device
US5790104A (en) 1996-06-25 1998-08-04 International Business Machines Corporation Multiple, moveable, customizable virtual pointing devices
US5933134A (en) 1996-06-25 1999-08-03 International Business Machines Corporation Touch screen virtual pointing device which goes into a translucent hibernation state when not in use
US6054990A (en) 1996-07-05 2000-04-25 Tran; Bao Q. Computer system with handwriting annotation
JPH1039748A (en) 1996-07-18 1998-02-13 Koonet:Kk Electronic map system, information providing device, and storage medium using it
US6002808A (en) 1996-07-26 1999-12-14 Mitsubishi Electric Information Technology Center America, Inc. Hand gesture control system
US6288707B1 (en) 1996-07-29 2001-09-11 Harald Philipp Capacitive position sensor
KR100260760B1 (en) 1996-07-31 2000-07-01 모리 하루오 Information display system with touch panel
US5943044A (en) 1996-08-05 1999-08-24 Interlink Electronics Force sensing semiconductive touchpad
US5818451A (en) 1996-08-12 1998-10-06 International Busienss Machines Corporation Computer programmed soft keyboard system, method and apparatus having user input displacement
US6208329B1 (en) 1996-08-13 2001-03-27 Lsi Logic Corporation Supplemental mouse button emulation system, method and apparatus for a coordinate based data input device
US5745116A (en) 1996-09-09 1998-04-28 Motorola, Inc. Intuitive gesture-based graphical user interface
US6308144B1 (en) 1996-09-26 2001-10-23 Computervision Corporation Method and apparatus for providing three-dimensional model associativity
GB9620464D0 (en) 1996-10-01 1996-11-20 Philips Electronics Nv Hand held image display device
US6650319B1 (en) 1996-10-29 2003-11-18 Elo Touchsystems, Inc. Touch screen based topological mapping with resistance framing design
US5854625A (en) 1996-11-06 1998-12-29 Synaptics, Incorporated Force sensing touchpad
US5870091A (en) 1996-11-07 1999-02-09 Adobe Systems Incorporated Combining palettes on a computer display
US6031469A (en) 1996-11-12 2000-02-29 Dodd; Jerry Ergonomic computer keyboard
US5883619A (en) 1996-11-12 1999-03-16 Primax Electronics Ltd. Computer mouse for scrolling a view of an image
US5748269A (en) 1996-11-21 1998-05-05 Westinghouse Air Brake Company Environmentally-sealed, convectively-cooled active matrix liquid crystal display (LCD)
US6154201A (en) 1996-11-26 2000-11-28 Immersion Corporation Control knob with multiple degrees of freedom and force feedback
JPH10171583A (en) 1996-12-05 1998-06-26 Wacom Co Ltd Position detector and position indicator therefor
JPH10171600A (en) 1996-12-06 1998-06-26 Brother Ind Ltd Input device
US6061177A (en) 1996-12-19 2000-05-09 Fujimoto; Kenneth Noboru Integrated computer display and graphical input apparatus and method
DE69626208T2 (en) 1996-12-20 2003-11-13 Hitachi Europe Ltd., Maidenhead Method and system for recognizing hand gestures
US6278441B1 (en) 1997-01-09 2001-08-21 Virtouch, Ltd. Tactile interface system for electronic data display system
US5850218A (en) 1997-02-19 1998-12-15 Time Warner Entertainment Company L.P. Inter-active program guide with default selection control
JPH10232734A (en) 1997-02-20 1998-09-02 Sharp Corp Virtual keyboard
US6105419A (en) 1997-03-25 2000-08-22 Recot, Inc. Apparatus and process for inspecting sealed packages
AU6792898A (en) 1997-03-31 1998-10-22 Ttsi (Technical Telephone Systems, Inc.) Computer-telephony integration employing an intelligent keyboard and method for same
US5943053A (en) 1997-04-01 1999-08-24 Sun Microsystems, Inc. Method and apparatus for expanding and contracting a window panel
US6073036A (en) 1997-04-28 2000-06-06 Nokia Mobile Phones Limited Mobile station with touch input having automatic symbol magnification function
US6049326A (en) 1997-05-12 2000-04-11 Siemens Information And Communication Networks, Inc. System and method for dual browser modes
US6054984A (en) 1997-05-27 2000-04-25 Hewlett-Packard Company System and method for graphically manipulating display elements in a signal measurement system
US5949345A (en) 1997-05-27 1999-09-07 Microsoft Corporation Displaying computer information to a driver of a vehicle
US6803905B1 (en) 1997-05-30 2004-10-12 International Business Machines Corporation Touch sensitive apparatus and method for improved visual feedback
US6289326B1 (en) 1997-06-04 2001-09-11 Lafleur Bernard B. Portable interactive kiosk
CH691711A5 (en) 1997-06-11 2001-09-14 Asulab Sa Watch with tactile reading and adjustment of horological functions.
US5910800A (en) 1997-06-11 1999-06-08 Microsoft Corporation Usage tips for on-screen touch-sensitive controls
JP3968477B2 (en) 1997-07-07 2007-08-29 ソニー株式会社 Information input device and information input method
US7002749B2 (en) 1997-07-08 2006-02-21 Kremen Stanley H Modular integral magnifier
US5974541A (en) 1997-07-22 1999-10-26 National Instruments Corporation GPIB system and method which provides asynchronous event notification
JPH1153093A (en) 1997-08-04 1999-02-26 Hitachi Ltd Input device
US5943052A (en) 1997-08-12 1999-08-24 Synaptics, Incorporated Method and apparatus for scroll bar control
JP3000972B2 (en) 1997-08-18 2000-01-17 日本電気株式会社 Information providing apparatus and machine-readable recording medium recording program
US6920619B1 (en) 1997-08-28 2005-07-19 Slavoljub Milekic User interface for removing an object from a display
JPH1173271A (en) 1997-08-28 1999-03-16 Sharp Corp Instructing device and processor and storage medium
JPH1178369A (en) 1997-09-03 1999-03-23 Plus Kogyo Kk Display system
JPH1185380A (en) 1997-09-10 1999-03-30 Seiko Denshi Kiki Kk Coordinate reader, its interface device, and coordinate reading system
US6104384A (en) 1997-09-12 2000-08-15 Ericsson, Inc. Image based keyboard for a small computing device
US6433801B1 (en) 1997-09-26 2002-08-13 Ericsson Inc. Method and apparatus for using a touch screen display on a portable intelligent communications device
US6084576A (en) 1997-09-27 2000-07-04 Leu; Neng-Chyang User friendly keyboard
US6037882A (en) 1997-09-30 2000-03-14 Levy; David H. Method and apparatus for inputting data to an electronic system
US6072494A (en) 1997-10-15 2000-06-06 Electric Planet, Inc. Method and apparatus for real-time gesture recognition
JPH11119911A (en) 1997-10-15 1999-04-30 Fujitsu Ten Ltd Switching device
GB2330670B (en) 1997-10-24 2002-09-11 Sony Uk Ltd Data processing
GB2330750B (en) 1997-10-24 2002-09-04 Sony Uk Ltd Audio processing
GB2330752B (en) 1997-10-24 2002-09-04 Sony Uk Ltd Audio processing
JPH11133816A (en) 1997-10-27 1999-05-21 Canon Inc Image forming device, method for inputting set information, and storage medium
US5995101A (en) 1997-10-29 1999-11-30 Adobe Systems Incorporated Multi-level tool tip
JP3794180B2 (en) 1997-11-11 2006-07-05 セイコーエプソン株式会社 Coordinate input system and coordinate input device
US6057845A (en) 1997-11-14 2000-05-02 Sensiva, Inc. System, method, and apparatus for generation and recognizing universal commands
US6400379B1 (en) 1997-11-25 2002-06-04 Pioneer Digital Technologies, Inc. Method and apparatus for selectively displaying additional information relating to broadcast information
SE510949C2 (en) 1997-11-25 1999-07-12 Saab Automobile Control Panel
SE9704330D0 (en) 1997-11-25 1997-11-25 Siemens Elema Ab Device Panel
FI104928B (en) 1997-11-27 2000-04-28 Nokia Mobile Phones Ltd Wireless Communication and a Method of Making a Wireless Communication Device
US6037937A (en) 1997-12-04 2000-03-14 Nortel Networks Corporation Navigation tool for graphical user interface
US6310610B1 (en) 1997-12-04 2001-10-30 Nortel Networks Limited Intelligent touch display
GB2332293A (en) 1997-12-11 1999-06-16 British Telecomm An Input Device
US6075531A (en) 1997-12-15 2000-06-13 International Business Machines Corporation Computer system and method of manipulating multiple graphical user interface components on a computer display with a proximity pointer
JPH11184669A (en) 1997-12-24 1999-07-09 Sharp Corp Information processing device and method therefor, and medium storing information processing device control program
US6757001B2 (en) 1999-03-30 2004-06-29 Research Investment Network, Inc. Method of using physical buttons in association with a display to access and execute functions available through associated hardware and software
US5933141A (en) 1998-01-05 1999-08-03 Gateway 2000, Inc. Mutatably transparent displays
JPH11194863A (en) 1998-01-06 1999-07-21 Poseidon Technical Systems:Kk Touch input detecting method and touch input detector
US7840912B2 (en) 2006-01-30 2010-11-23 Apple Inc. Multi-touch gesture dictionary
US8479122B2 (en) 2004-07-30 2013-07-02 Apple Inc. Gestures for touch sensitive input devices
US9239673B2 (en) 1998-01-26 2016-01-19 Apple Inc. Gesturing with a multipoint sensing device
US7614008B2 (en) 2004-07-30 2009-11-03 Apple Inc. Operation of a computer with touch screen interface
US20060033724A1 (en) 2004-07-30 2006-02-16 Apple Computer, Inc. Virtual input device placement on a touch screen user interface
US7844914B2 (en) 2004-07-30 2010-11-30 Apple Inc. Activating virtual keys of a touch-screen virtual keyboard
US20070177804A1 (en) 2006-01-30 2007-08-02 Apple Computer, Inc. Multi-touch gesture dictionary
US7663607B2 (en) 2004-05-06 2010-02-16 Apple Inc. Multipoint touchscreen
US7760187B2 (en) 2004-07-30 2010-07-20 Apple Inc. Visual expander
KR100595924B1 (en) 1998-01-26 2006-07-05 웨인 웨스터만 Method and apparatus for integrating manual input
US9292111B2 (en) 1998-01-26 2016-03-22 Apple Inc. Gesturing with a multipoint sensing device
US7800592B2 (en) 2005-03-04 2010-09-21 Apple Inc. Hand held electronic device with multiple touch sensing devices
JP2938420B2 (en) 1998-01-30 1999-08-23 インターナショナル・ビジネス・マシーンズ・コーポレイション Function selection method and apparatus, storage medium storing control program for selecting functions, object operation method and apparatus, storage medium storing control program for operating objects, storage medium storing composite icon
US6175610B1 (en) 1998-02-11 2001-01-16 Siemens Aktiengesellschaft Medical technical system controlled by vision-detected operator activity
US6249606B1 (en) 1998-02-19 2001-06-19 Mindmaker, Inc. Method and system for gesture category recognition and training using a feature vector
US6380853B1 (en) 1998-02-23 2002-04-30 Marconi Commerce Systems Inc. Customer-sensitive dispenser using proximity sensing devices
US6013956A (en) 1998-02-23 2000-01-11 Cooper Automotive Products, Inc. Touch control switches for vehicles
US6639577B2 (en) 1998-03-04 2003-10-28 Gemstar-Tv Guide International, Inc. Portable information display device with ergonomic bezel
US6198515B1 (en) 1998-03-16 2001-03-06 Joseph W. Cole Apparatus and method for controlled sealing between bezel and CRT
US6130665A (en) 1998-04-01 2000-10-10 Telefonaktiebolaget Lm Ericsson Touch screen handling
US6313853B1 (en) 1998-04-16 2001-11-06 Nortel Networks Limited Multi-service user interface
US5956291A (en) 1998-04-17 1999-09-21 Ductech, Llc Underwater diving assistant apparatus
US6211856B1 (en) 1998-04-17 2001-04-03 Sung M. Choi Graphical user interface touch screen with an auto zoom feature
US6278443B1 (en) 1998-04-30 2001-08-21 International Business Machines Corporation Touch screen with random finger placement and rolling on screen to control the movement of information on-screen
US6057540A (en) 1998-04-30 2000-05-02 Hewlett-Packard Co Mouseless optical and position translation type screen pointer control for a computer system
US6050825A (en) 1998-05-08 2000-04-18 Nichol; Claudia D. Opaque, one-size-fits-all computer keyboard cover which covers only the three or four alpha-numeric rows
US6473102B1 (en) 1998-05-11 2002-10-29 Apple Computer, Inc. Method and system for automatically resizing and repositioning windows in response to changes in display
KR100327209B1 (en) 1998-05-12 2002-04-17 윤종용 Software keyboard system using the drawing of stylus and method for recognizing keycode therefor
US5977867A (en) 1998-05-29 1999-11-02 Nortel Networks Corporation Touch pad panel with tactile feedback
JP4033582B2 (en) 1998-06-09 2008-01-16 株式会社リコー Coordinate input / detection device and electronic blackboard system
JP2000010705A (en) 1998-06-17 2000-01-14 Nec Corp Output image deforming system and record medium recording program for output image deformation
US6429846B2 (en) 1998-06-23 2002-08-06 Immersion Corporation Haptic feedback for touchpads and other touch controls
JP3792920B2 (en) 1998-12-25 2006-07-05 株式会社東海理化電機製作所 Touch operation input device
US6347290B1 (en) 1998-06-24 2002-02-12 Compaq Information Technologies Group, L.P. Apparatus and method for detecting and executing positional and gesture commands corresponding to movement of handheld computing device
JP3420504B2 (en) 1998-06-30 2003-06-23 キヤノン株式会社 Information processing method
US6131299A (en) 1998-07-01 2000-10-17 Faro Technologies, Inc. Display device for a coordinate measurement machine
US6188391B1 (en) 1998-07-09 2001-02-13 Synaptics, Inc. Two-layer capacitive touchpad and method of making same
US7831930B2 (en) 2001-11-20 2010-11-09 Universal Electronics Inc. System and method for displaying a user interface for a remote control application
US6135958A (en) 1998-08-06 2000-10-24 Acuson Corporation Ultrasound imaging system with touch-pad pointing device
US6169538B1 (en) 1998-08-13 2001-01-02 Motorola, Inc. Method and apparatus for implementing a graphical user interface keyboard and a text buffer on electronic devices
US7079114B1 (en) 1998-08-21 2006-07-18 Peter Smith Interactive methods for design of automobiles
US6326956B1 (en) 1998-08-24 2001-12-04 Intertactile Technologies Corporation Circuit control devices utilizing electronic display screen light
US6271835B1 (en) 1998-09-03 2001-08-07 Nortel Networks Limited Touch-screen input device
US6359572B1 (en) 1998-09-03 2002-03-19 Microsoft Corporation Dynamic keyboard
AU5808299A (en) 1998-09-04 2000-03-27 Innovative Solutions And Support Inc. Flat panel display using dual cpu's for an aircraft cockpit
US6333753B1 (en) 1998-09-14 2001-12-25 Microsoft Corporation Technique for implementing an on-demand display widget through controlled fading initiated by user contact with a touch sensitive input device
US7256770B2 (en) 1998-09-14 2007-08-14 Microsoft Corporation Method for displaying information responsive to sensing a physical presence proximate to a computer input device
US20020018051A1 (en) 1998-09-15 2002-02-14 Mona Singh Apparatus and method for moving objects on a touchscreen display
US6501515B1 (en) 1998-10-13 2002-12-31 Sony Corporation Remote control system
US6292178B1 (en) 1998-10-19 2001-09-18 Allergan Sales, Inc. Screen navigation control apparatus for ophthalmic surgical instruments
JP2000132305A (en) 1998-10-23 2000-05-12 Olympus Optical Co Ltd Operation input device
US6229502B1 (en) 1998-11-03 2001-05-08 Cylark Development Llc Electronic book
US6874023B1 (en) 1998-11-10 2005-03-29 Claria Corporation Web based email control center for monitoring and providing a sumary of the detected event information organized according to relationships between the user and network sites
JP2000163193A (en) 1998-11-25 2000-06-16 Seiko Epson Corp Portable information equipment and information storage medium
JP2000163444A (en) 1998-11-25 2000-06-16 Seiko Epson Corp Portable information device and information storage medium
US6466036B1 (en) 1998-11-25 2002-10-15 Harald Philipp Charge transfer capacitance measurement circuit
JP4542637B2 (en) 1998-11-25 2010-09-15 セイコーエプソン株式会社 Portable information device and information storage medium
US6154194A (en) 1998-12-03 2000-11-28 Ericsson Inc. Device having adjustable touch-based display of data
US6222465B1 (en) 1998-12-09 2001-04-24 Lucent Technologies Inc. Gesture-based computer interface
US6246395B1 (en) 1998-12-17 2001-06-12 Hewlett-Packard Company Palm pressure rejection method and apparatus for touchscreens
US6259436B1 (en) 1998-12-22 2001-07-10 Ericsson Inc. Apparatus and method for determining selection of touchable items on a computer touchscreen by an imprecise touch
US6643824B1 (en) 1999-01-15 2003-11-04 International Business Machines Corporation Touch screen region assist for hypertext links
US6535200B2 (en) 1999-01-25 2003-03-18 Harald Philipp Capacitive position sensor
ATE517426T1 (en) 1999-01-26 2011-08-15 Limited Qrg CAPACITIVE TRANSDUCER AND ARRANGEMENT
US6246862B1 (en) 1999-02-03 2001-06-12 Motorola, Inc. Sensor controlled user interface for portable communication device
JP2000231670A (en) 1999-02-10 2000-08-22 Sharp Corp Vending management device with touch panel
JP2000250697A (en) 1999-03-03 2000-09-14 Yazaki Corp Function display method for operating button of multifonctional switch device and multifunctional switch device
US6295049B1 (en) 1999-03-03 2001-09-25 Richard T. Minner Computer system utilizing graphical user interface with hysteresis to inhibit accidental selection of a region due to unintended cursor motion and method
US7293231B1 (en) 1999-03-18 2007-11-06 British Columbia Ltd. Data entry for personal computing devices
US6545670B1 (en) 1999-05-11 2003-04-08 Timothy R. Pryor Methods and apparatus for man machine interfaces and related activity
US6710771B1 (en) 1999-05-13 2004-03-23 Sony Corporation Information processing method and apparatus and medium
US6920500B2 (en) 1999-05-20 2005-07-19 Prosia Overseas, Inc. Method and apparatus for providing simplified access to the internet
EP1054321A3 (en) 1999-05-21 2002-06-19 Sony Corporation Information processing method and apparatus
US7030863B2 (en) 2000-05-26 2006-04-18 America Online, Incorporated Virtual keyboard system with automatic correction
US7151528B2 (en) 1999-06-22 2006-12-19 Cirque Corporation System for disposing a proximity sensitive touchpad behind a mobile phone keypad
US6525711B1 (en) 1999-06-24 2003-02-25 Interval Research Corp. Haptic interface including clutch control
US6639584B1 (en) 1999-07-06 2003-10-28 Chuang Li Methods and apparatus for controlling a portable electronic device using a touchpad
US6904405B2 (en) 1999-07-17 2005-06-07 Edwin A. Suominen Message recognition using shared language model
US6337678B1 (en) 1999-07-21 2002-01-08 Tactiva Incorporated Force feedback computer input and output device with coordinated haptic elements
US6677934B1 (en) 1999-07-30 2004-01-13 L-3 Communications Infrared touch panel with improved sunlight rejection
US6489978B1 (en) 1999-08-06 2002-12-03 International Business Machines Corporation Extending the opening time of state menu items for conformations of multiple changes
US20020173721A1 (en) 1999-08-20 2002-11-21 Novasonics, Inc. User interface for handheld imaging devices
GB9920301D0 (en) 1999-08-27 1999-11-03 Philipp Harald Level sensing
US7149893B1 (en) 1999-09-07 2006-12-12 Poofaway.Com, Inc. System and method for enabling the originator of an electronic mail message to preset an expiration time, date, and/or event, and to control processing or handling by a recipient
US6377009B1 (en) 1999-09-08 2002-04-23 Harald Philipp Capacitive closure obstruction sensor
US6411287B1 (en) 1999-09-08 2002-06-25 Elo Touchsystems, Inc. Stress seal for acoustic wave touchscreens
US6424338B1 (en) 1999-09-30 2002-07-23 Gateway, Inc. Speed zone touchpad
GB9924177D0 (en) 1999-10-12 1999-12-15 Srs Technology Limited Communication and control system
EP1094383A3 (en) 1999-10-18 2005-11-02 Honeywell International, Inc. Methods and apparatus for cursor control
JP2001134382A (en) 1999-11-04 2001-05-18 Sony Corp Graphic processor
US20030132950A1 (en) 2001-11-27 2003-07-17 Fahri Surucu Detecting, classifying, and interpreting input events based on stimuli in multiple sensory domains
US6757002B1 (en) 1999-11-04 2004-06-29 Hewlett-Packard Development Company, L.P. Track pad pointing device with areas of specialized function
US8482535B2 (en) 1999-11-08 2013-07-09 Apple Inc. Programmable tactile touch screen displays and man-machine interfaces for improved vehicle instrumentation and telematics
US20020196227A1 (en) 1999-11-15 2002-12-26 Samuel Surloff Method and apparatus for providing simplified access to the internet
US6677933B1 (en) 1999-11-15 2004-01-13 Espial Group Inc. Method and apparatus for operating a virtual keyboard
JP2001147918A (en) 1999-11-24 2001-05-29 Casio Comput Co Ltd Information display device and storage medium with stored program for information display processing
US6920607B1 (en) 1999-12-15 2005-07-19 Microsoft Corp. Methods and systems for dynamically creating user interfaces
US6414674B1 (en) 1999-12-17 2002-07-02 International Business Machines Corporation Data processing system and method including an I/O touch pad having dynamically alterable location indicators
US7434177B1 (en) 1999-12-20 2008-10-07 Apple Inc. User interface for providing consolidation and access
US6573844B1 (en) 2000-01-18 2003-06-03 Microsoft Corporation Predictive keyboard
US6654733B1 (en) 2000-01-18 2003-11-25 Microsoft Corporation Fuzzy keyboard
US6661920B1 (en) 2000-01-19 2003-12-09 Palm Inc. Method and apparatus for multiple simultaneously active data entry mechanisms on a computer system
US20020140633A1 (en) 2000-02-03 2002-10-03 Canesta, Inc. Method and system to present immersion virtual simulations using three-dimensional measurement
JP2001230992A (en) 2000-02-18 2001-08-24 Nidec Copal Corp Image display device and its control method
US6882641B1 (en) 2000-02-23 2005-04-19 Lucent Technologies Inc. Call center queue administration
US6798768B1 (en) 2000-02-23 2004-09-28 Lucent Technologies Inc. Multimedia call routing in an IP network
US8712792B2 (en) 2000-02-24 2014-04-29 Webmd, Llc Personalized health communication system
US6597345B2 (en) 2000-03-03 2003-07-22 Jetway Technologies Ltd. Multifunctional keypad on touch screen
JP3812274B2 (en) 2000-03-17 2006-08-23 株式会社日立製作所 VIDEO ACCESS DEVICE AND RECORDING MEDIUM RECORDING VIDEO ACCESS PROGRAM
JP2001265481A (en) 2000-03-21 2001-09-28 Nec Corp Method and device for displaying page information and storage medium with program for displaying page information stored
US6791467B1 (en) 2000-03-23 2004-09-14 Flextronics Semiconductor, Inc. Adaptive remote controller
US6543684B1 (en) 2000-03-28 2003-04-08 Ncr Corporation Transaction terminal with privacy shield for touch-screen pin entry
US6456952B1 (en) 2000-03-29 2002-09-24 Ncr Coporation System and method for touch screen environmental calibration
US6624833B1 (en) 2000-04-17 2003-09-23 Lucent Technologies Inc. Gesture-based input interface system with shadow detection
US6570584B1 (en) 2000-05-15 2003-05-27 Eastman Kodak Company Broad color gamut display
US6611252B1 (en) 2000-05-17 2003-08-26 Dufaux Douglas P. Virtual data input device
US20020044161A1 (en) 2000-05-22 2002-04-18 Kazuaki Sugai Multi-window display system and method for displaying video data and storage medium
US20020075317A1 (en) 2000-05-26 2002-06-20 Dardick Technologies System and method for an on-demand script-activated virtual keyboard
IL136652A0 (en) 2000-06-08 2001-06-14 Arlinsky David A closed-loop control system in a car
KR20010111166A (en) 2000-06-08 2001-12-17 윤종용 A method of changing menu icon and editing menu configuration for a mobile telephone
FI20001429A (en) 2000-06-15 2001-12-16 Nokia Corp Choosing an alternative
US6442440B1 (en) 2000-06-24 2002-08-27 Dade Behring Inc. Computer interface module having a flat menu
US6803906B1 (en) 2000-07-05 2004-10-12 Smart Technologies, Inc. Passive touch system and method of detecting user input
US8287374B2 (en) 2000-07-07 2012-10-16 Pryor Timothy R Reconfigurable control displays for games, toys, and other applications
US7466843B2 (en) 2000-07-07 2008-12-16 Pryor Timothy R Multi-functional control and entertainment systems
JP4543513B2 (en) 2000-07-17 2010-09-15 ソニー株式会社 Bidirectional communication system, display device, base device, and bidirectional communication method
US6944591B1 (en) 2000-07-27 2005-09-13 International Business Machines Corporation Audio support system for controlling an e-mail system in a remote computer
EP1311938B1 (en) 2000-08-17 2009-04-01 John Molgaard A graphical user interface for data entry
JP4197220B2 (en) 2000-08-17 2008-12-17 アルパイン株式会社 Operating device
DE10042300A1 (en) 2000-08-29 2002-03-28 Axel C Burgbacher Electronic musical instrument with tone generator contg. input members
US6480188B1 (en) 2000-09-06 2002-11-12 Digital On-Demand Thumbwheel selection system
JP2002157078A (en) 2000-09-07 2002-05-31 Sony Corp Processor and method for information processing, and recording medium
US6680677B1 (en) 2000-10-06 2004-01-20 Logitech Europe S.A. Proximity detector to indicate function of a key
US6421234B1 (en) 2000-10-10 2002-07-16 Juniper Systems, Inc. Handheld electronics device having ergonomic features
US6942571B1 (en) 2000-10-16 2005-09-13 Bally Gaming, Inc. Gaming device with directional and speed control of mechanical reels using touch screen
JP4284855B2 (en) 2000-10-25 2009-06-24 ソニー株式会社 Information input / output system, information input / output method, and program storage medium
TW486657B (en) 2000-10-26 2002-05-11 Animeta Systems Inc Browser interface operation device and its browsing method
US6501464B1 (en) 2000-10-31 2002-12-31 Intel Corporation On-screen transparent keyboard interface
US6593916B1 (en) 2000-11-03 2003-07-15 James L. Aroyan Touchscreen having multiple parallel connections to each electrode in a series resistor chain on the periphery of the touch area
WO2002039245A2 (en) 2000-11-09 2002-05-16 Change Tools, Inc. A user definable interface system, method and computer program product
US6903730B2 (en) 2000-11-10 2005-06-07 Microsoft Corporation In-air gestures for electromagnetic coordinate digitizers
US7319454B2 (en) 2000-11-10 2008-01-15 Microsoft Corporation Two-button mouse input using a stylus
US6703999B1 (en) 2000-11-13 2004-03-09 Toyota Jidosha Kabushiki Kaisha System for computer user interface
DE10059906A1 (en) 2000-12-01 2002-06-06 Bs Biometric Systems Gmbh Pressure-sensitive surface for use with a screen or a display linked to a computer displays fields sensitive to touch pressure for triggering a computer program function related to the appropriate field.
US20040036622A1 (en) 2000-12-15 2004-02-26 Semyon Dukach Apparatuses, methods, and computer programs for displaying information on signs
US7190348B2 (en) 2000-12-26 2007-03-13 International Business Machines Corporation Method for touchscreen data input
US6457834B1 (en) 2001-01-24 2002-10-01 Scram Technologies, Inc. Optical system for display panel
US6677932B1 (en) 2001-01-28 2004-01-13 Finger Works, Inc. System and method for recognizing touch typing under limited tactile feedback conditions
US20020135615A1 (en) 2001-01-31 2002-09-26 Microsoft Corporation Overlaid display for electronic devices
US6670894B2 (en) 2001-02-05 2003-12-30 Carsten Mehring System and method for keyboard independent touch typing
US6570557B1 (en) 2001-02-10 2003-05-27 Finger Works, Inc. Multi-touch system and method for emulating modifier keys via fingertip chords
US7030861B1 (en) 2001-02-10 2006-04-18 Wayne Carl Westerman System and method for packing multi-touch gestures onto a hand
JP2002244781A (en) 2001-02-15 2002-08-30 Wacom Co Ltd Input system, program, and recording medium
US20070083823A1 (en) 2001-02-15 2007-04-12 Denny Jaeger Scroll bar for computer display
US20080088587A1 (en) 2001-02-22 2008-04-17 Timothy Pryor Compact rtd instrument panels and computer interfaces
US20080024463A1 (en) 2001-02-22 2008-01-31 Timothy Pryor Reconfigurable tactile control display applications
US7133531B2 (en) 2001-02-27 2006-11-07 Nissim Karpenstein Device using analog controls to mix compressed digital audio data
JP2002259036A (en) 2001-02-28 2002-09-13 Sony Corp Information processor, information processing method, recording medium, and program
US7904358B2 (en) 2001-02-28 2011-03-08 Goldman Sachs & Co. Computerized interface for monitoring financial information and executing financial transactions
US6806869B2 (en) 2001-03-02 2004-10-19 Seiko Epson Corporation Data processing system utilizing discrete operating device
US6543947B2 (en) 2001-03-14 2003-04-08 Lee Jose H. Keyboard having keys arranged in a pan configuration
US6621483B2 (en) 2001-03-16 2003-09-16 Agilent Technologies, Inc. Optical screen pointing device with inertial properties
GB2373596B (en) 2001-03-19 2003-02-12 Synigence Plc Apparatus for facilitating access to information
US7012595B2 (en) 2001-03-30 2006-03-14 Koninklijke Philips Electronics N.V. Handheld electronic device with touch pad
US6724366B2 (en) 2001-04-03 2004-04-20 Peter James Crawford Thumb actuated x-y input device
JP3597141B2 (en) 2001-04-03 2004-12-02 泰鈞 温 Information input device and method, mobile phone and character input method of mobile phone
TW504916B (en) 2001-04-24 2002-10-01 Inventec Appliances Corp Method capable of generating different input values by pressing a single key from multiple directions
US7168046B2 (en) 2001-04-26 2007-01-23 Lg Electronics Inc. Method and apparatus for assisting data input to a portable information terminal
US6965375B1 (en) 2001-04-27 2005-11-15 Palm, Inc. Compact integrated touch panel display for a handheld device
EP1393189A4 (en) 2001-05-02 2007-06-13 Bitstream Inc Methods, systems, and programming for displaying media scaled-down by a variable scale factor
US20050024341A1 (en) 2001-05-16 2005-02-03 Synaptics, Inc. Touch screen with user interface enhancement
US7730401B2 (en) 2001-05-16 2010-06-01 Synaptics Incorporated Touch screen with user interface enhancement
JP3800984B2 (en) 2001-05-21 2006-07-26 ソニー株式会社 User input device
US6992659B2 (en) 2001-05-22 2006-01-31 Palmone, Inc. High transparency integrated enclosure touch screen assembly for a portable hand held device
TW521205B (en) 2001-06-05 2003-02-21 Compal Electronics Inc Touch screen capable of controlling amplification with pressure
US6904570B2 (en) 2001-06-07 2005-06-07 Synaptics, Inc. Method and apparatus for controlling a display of data on a display screen
US7308652B2 (en) 2001-06-08 2007-12-11 Lenovo Singapore Pte. Ltd Entry of a password through a touch-sensitive computer screen
US6975304B1 (en) 2001-06-11 2005-12-13 Handspring, Inc. Interface for processing of an alternate symbol in a computer device
US6583676B2 (en) 2001-06-20 2003-06-24 Apple Computer, Inc. Proximity/touch detector and calibration circuit
JP2003005912A (en) 2001-06-20 2003-01-10 Hitachi Ltd Display device with touch panel and display method
US7512964B2 (en) 2001-06-29 2009-03-31 Cisco Technology System and method for archiving multiple downloaded recordable media content
FI116591B (en) 2001-06-29 2005-12-30 Nokia Corp Method and apparatus for performing a function
US20030001899A1 (en) 2001-06-29 2003-01-02 Nokia Corporation Semi-transparent handwriting recognition UI
US6670951B2 (en) 2001-07-03 2003-12-30 Hewlett-Packard Development Company, L.P. Methods and systems for increasing the input efficiency of personal digital assistants and other handheld stylus-engagable computing devices
US20050134578A1 (en) 2001-07-13 2005-06-23 Universal Electronics Inc. System and methods for interacting with a control environment
US6478432B1 (en) 2001-07-13 2002-11-12 Chad D. Dyner Dynamically generated interactive real imaging device
US6961912B2 (en) 2001-07-18 2005-11-01 Xerox Corporation Feedback mechanism for use with visual selection methods
US20030030664A1 (en) 2001-08-13 2003-02-13 Parry Travis J. Customizable control panel software
DE10140874A1 (en) 2001-08-21 2003-03-20 Dosch & Amand Gmbh & Co Kg Touch screen type graphical user interface in which the sizing of screen areas is automatically adjusted to simplify user input and make control of devices, especially communications devices more user friendly
US6972749B2 (en) 2001-08-29 2005-12-06 Microsoft Corporation Touch-sensitive device for scrolling a document on a display
JP4176017B2 (en) 2001-09-21 2008-11-05 インターナショナル・ビジネス・マシーンズ・コーポレーション INPUT DEVICE, COMPUTER DEVICE, INPUT OBJECT IDENTIFICATION METHOD, AND COMPUTER PROGRAM
JP2003173237A (en) 2001-09-28 2003-06-20 Ricoh Co Ltd Information input-output system, program and storage medium
US7254775B2 (en) 2001-10-03 2007-08-07 3M Innovative Properties Company Touch panel system and method for distinguishing multiple touch inputs
GB2380583A (en) 2001-10-04 2003-04-09 Ilam Samson Touch pad/screen for electronic equipment
IL161138A0 (en) 2001-10-04 2004-08-31 Xtl Biopharmaceuticals Ltd Treatment of hepatitis b virus infection with human monoclonal antibodies
US20030071850A1 (en) 2001-10-12 2003-04-17 Microsoft Corporation In-place adaptive handwriting input method and system
US6728032B2 (en) 2001-10-17 2004-04-27 Infocus Corporation Rear projection display system
US8117565B2 (en) 2001-10-18 2012-02-14 Viaclix, Inc. Digital image magnification for internet appliance
US7345671B2 (en) 2001-10-22 2008-03-18 Apple Inc. Method and apparatus for use of rotational user inputs
US7046230B2 (en) 2001-10-22 2006-05-16 Apple Computer, Inc. Touch pad handheld device
US7084856B2 (en) 2001-10-22 2006-08-01 Apple Computer, Inc. Mouse having a rotary dial
US7312785B2 (en) 2001-10-22 2007-12-25 Apple Inc. Method and apparatus for accelerated scrolling
GB2382292A (en) 2001-11-17 2003-05-21 Oliver Wallington Digital display with scrolling means
US7009599B2 (en) 2001-11-20 2006-03-07 Nokia Corporation Form factor for portable device
US6762752B2 (en) 2001-11-29 2004-07-13 N-Trig Ltd. Dual function input device and method
US6938221B2 (en) 2001-11-30 2005-08-30 Microsoft Corporation User interface for stylus-based user input
US7083342B2 (en) 2001-12-21 2006-08-01 Griffin Jason T Keyboard arrangement
DE10251296A1 (en) 2002-11-03 2004-05-19 Trachte, Ralf, Dr. Device for computer input has a touch-sensitive input surface with multiple input surface areas and a control unit coupled to the input surface.
US6690387B2 (en) 2001-12-28 2004-02-10 Koninklijke Philips Electronics N.V. Touch-screen image scrolling system and method
US6952203B2 (en) 2002-01-08 2005-10-04 International Business Machines Corporation Touchscreen user interface: Bluetooth™ stylus for performing right mouse clicks
US6977643B2 (en) 2002-01-10 2005-12-20 International Business Machines Corporation System and method implementing non-physical pointers for computer devices
US20030197736A1 (en) 2002-01-16 2003-10-23 Murphy Michael W. User interface for character entry using a minimum number of selection keys
US20030132922A1 (en) 2002-01-17 2003-07-17 Harald Philipp Touch screen detection apparatus
GB0201074D0 (en) 2002-01-18 2002-03-06 3G Lab Ltd Graphic user interface for data processing device
US7075512B1 (en) 2002-02-07 2006-07-11 Palmsource, Inc. Method and system for navigating a display screen for locating a desired item of information
AUPS058602A0 (en) 2002-02-15 2002-03-14 Canon Kabushiki Kaisha Representing a plurality of independent data items
AU2003217587A1 (en) 2002-02-15 2003-09-09 Canesta, Inc. Gesture recognition system using depth perceptive sensors
EP1479065A4 (en) 2002-02-26 2009-11-11 Cirque Corp Touchpad having fine and coarse input resolution
US6900795B1 (en) 2002-02-27 2005-05-31 Knight, Iii Jack Unitary molded lens filter for touch screen interface
US6832138B1 (en) 2002-02-28 2004-12-14 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
JP3847641B2 (en) 2002-02-28 2006-11-22 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, information processing program, computer-readable recording medium storing information processing program, and information processing method
US6985801B1 (en) 2002-02-28 2006-01-10 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
CA2478303C (en) 2002-03-08 2013-11-26 Fleettrakker, L.L.C. Equipment tracking system and method
JP4061274B2 (en) 2002-03-08 2008-03-12 レベレイションズ・イン・デザイン・リミテッド・パートナーシップ Electrical device controller
US7120872B2 (en) 2002-03-25 2006-10-10 Microsoft Corporation Organizing, editing, and rendering digital ink
US6927761B2 (en) 2002-03-29 2005-08-09 3M Innovative Properties Company Moisture deflector for capacitive NFI touch screens for use with bezels of conductive material
US7038659B2 (en) 2002-04-06 2006-05-02 Janusz Wiktor Rajkowski Symbol encoding apparatus and method
US6658994B1 (en) 2002-04-10 2003-12-09 Chromalox, Inc. Modular assembly for a holding cabinet controller
US7466307B2 (en) 2002-04-11 2008-12-16 Synaptics Incorporated Closed-loop sensor on a solid-state object position detector
US20030193481A1 (en) 2002-04-12 2003-10-16 Alexander Sokolsky Touch-sensitive input overlay for graphical user interface
JP2003316502A (en) 2002-04-25 2003-11-07 Sony Corp Terminal equipment and character input method
JP2003323259A (en) 2002-05-02 2003-11-14 Nec Corp Information processing apparatus
US7746325B2 (en) 2002-05-06 2010-06-29 3M Innovative Properties Company Method for improving positioned accuracy for a determined touch input
US7890865B2 (en) 2002-05-09 2011-02-15 Microsoft Corporation Methods and apparatuses for providing message information in graphical user interfaces based on user inputs
TWI238348B (en) 2002-05-13 2005-08-21 Kyocera Corp Portable information terminal, display control device, display control method, and recording media
US7051284B2 (en) 2002-05-16 2006-05-23 Microsoft Corporation Displaying information to indicate both the importance and the urgency of the information
WO2003098421A1 (en) 2002-05-16 2003-11-27 Sony Corporation Inputting method and inputting apparatus
JP3852368B2 (en) 2002-05-16 2006-11-29 ソニー株式会社 Input method and data processing apparatus
JP2003338954A (en) 2002-05-21 2003-11-28 Toshiba Corp Digital still camera
KR100941948B1 (en) 2002-05-21 2010-02-11 코닌클리케 필립스 일렉트로닉스 엔.브이. A system for selecting and entering objects and a method for entering objects from a set of objects and compuetr readable medium for storing software code for implementing the method
US7069515B1 (en) 2002-05-21 2006-06-27 Claria Corporation Method and apparatus for displaying messages in computer systems
US20030222977A1 (en) 2002-06-03 2003-12-04 Kazutora Yoshino Intelligent system and 3D virtual object generator
US7384158B2 (en) 2003-01-08 2008-06-10 Silicon Optix Inc Image projection system and method
FI20021162A0 (en) 2002-06-14 2002-06-14 Nokia Corp Electronic device and a method for administering its keypad
US7219308B2 (en) 2002-06-21 2007-05-15 Microsoft Corporation User interface for media player program
FI112119B (en) 2002-06-25 2003-10-31 Nokia Corp Touch screen control command interpreting method for electronic device e.g. mobile station, involves interpreting contact area larger than area before touch, as same area when area has been touched for release of touch
CA2490292A1 (en) 2002-06-26 2004-01-08 Vkb Inc. Multifunctional integrated image sensor and application to virtual interface technology
JP2004038896A (en) 2002-06-28 2004-02-05 Clarion Co Ltd Display control means
US7023427B2 (en) 2002-06-28 2006-04-04 Microsoft Corporation Method and system for detecting multiple touches on a touch-sensitive screen
US11275405B2 (en) 2005-03-04 2022-03-15 Apple Inc. Multi-functional hand-held device
US7656393B2 (en) 2005-03-04 2010-02-02 Apple Inc. Electronic device having display and surrounding touch sensitive bezel for user interface and control
US20040056849A1 (en) 2002-07-25 2004-03-25 Andrew Lohbihler Method and apparatus for powering, detecting and locating multiple touch input devices on a touch screen
US7058902B2 (en) 2002-07-30 2006-06-06 Microsoft Corporation Enhanced on-object context menus
JP4115198B2 (en) 2002-08-02 2008-07-09 株式会社日立製作所 Display device with touch panel
US6896375B2 (en) 2002-08-16 2005-05-24 Infocus Corporation Rear projection display device having multiple mirrors that are substantially parallel to a screen
US7151530B2 (en) 2002-08-20 2006-12-19 Canesta, Inc. System and method for determining an input selected by a user through a virtual interface
EP1394640B1 (en) 2002-08-30 2010-03-24 Asulab S.A. Timepiece with tactile reading and operation of time information
JP2004104594A (en) 2002-09-11 2004-04-02 Toshiba Corp Digital still camera and method for inputting user instruction
US7331868B2 (en) 2002-09-13 2008-02-19 Igt Wagering gaming device providing physical stimulation responses to various components of the gaming device
US7789756B2 (en) 2002-09-13 2010-09-07 Igt Wagering gaming device having simulated control of movement of game functional elements
JP2004110388A (en) 2002-09-18 2004-04-08 Sharp Corp Device having touch panel
GB0222094D0 (en) 2002-09-24 2002-10-30 Koninkl Philips Electronics Nv Graphical user interface navigation method and apparatus
JP2004118917A (en) 2002-09-25 2004-04-15 Clarion Co Ltd Electronic equipment and navigation apparatus
US8443045B2 (en) 2002-10-01 2013-05-14 Honda Motor Co., Ltd. Storage of selected e-mails including attachments in document management system
US20040080529A1 (en) 2002-10-24 2004-04-29 Wojcik Paul Kazimierz Method and system for securing text-entry in a web form over a computer network
US20040095311A1 (en) 2002-11-19 2004-05-20 Motorola, Inc. Body-centric virtual interactive apparatus and method
AU2003302554A1 (en) 2002-11-29 2004-06-23 Gnvention Ltd. Interface device for portable storage media
WO2004051392A2 (en) 2002-11-29 2004-06-17 Koninklijke Philips Electronics N.V. User interface with displaced representation of touch area
US7081887B2 (en) 2002-12-19 2006-07-25 Intel Corporation Method and apparatus for positioning a software keyboard
WO2004059610A1 (en) 2002-12-19 2004-07-15 Anders Trell Trust Computer input device
US7219829B2 (en) 2002-12-24 2007-05-22 Tracey Treat Carrier which is aesthetically irreversibly convertible from planar blank to closed package for coded card and methods for manufacture and use of the same
US6927763B2 (en) 2002-12-30 2005-08-09 Motorola, Inc. Method and system for providing a disambiguated keypad
US7898529B2 (en) 2003-01-08 2011-03-01 Autodesk, Inc. User interface having a placement and layout suitable for pen-based computers
US7337634B2 (en) 2003-01-09 2008-03-04 General Electric Company Washer/dryer touch sensitive graphical user interface
US7194699B2 (en) 2003-01-14 2007-03-20 Microsoft Corporation Animating images to reflect user selection
US7149981B1 (en) 2003-01-16 2006-12-12 Sprint Spectrum L.P. Method and system for facilitating selection of a portion of non-focusable object presented by a browser application
US7098896B2 (en) 2003-01-16 2006-08-29 Forword Input Inc. System and method for continuous stroke word-based text input
US7305129B2 (en) 2003-01-29 2007-12-04 Microsoft Corporation Methods and apparatus for populating electronic forms from scanned documents
US20040145601A1 (en) 2003-01-29 2004-07-29 International Business Machines Corporation Method and a device for providing additional functionality to a separate application
US6972401B2 (en) 2003-01-30 2005-12-06 Smart Technologies Inc. Illuminated bezel and touch system incorporating the same
US20040150669A1 (en) 2003-01-31 2004-08-05 Sabiers Mark L. Graphical user interface for describing the state of a managed system
US7158123B2 (en) 2003-01-31 2007-01-02 Xerox Corporation Secondary touch contextual sub-menu navigation for touch screen interface
EP1445922A1 (en) 2003-02-06 2004-08-11 Dialog Semiconductor GmbH Monolithic optical read-out circuit
US20040155888A1 (en) 2003-02-11 2004-08-12 Padgitt David Gary Method for displaying the contents of a collection of media objects
US20040160419A1 (en) 2003-02-11 2004-08-19 Terradigital Systems Llc. Method for entering alphanumeric characters into a graphical user interface
JP3844076B2 (en) 2003-03-07 2006-11-08 セイコーエプソン株式会社 Image processing system, projector, program, information storage medium, and image processing method
US7103852B2 (en) 2003-03-10 2006-09-05 International Business Machines Corporation Dynamic resizing of clickable areas of touch screen applications
US20040183833A1 (en) 2003-03-19 2004-09-23 Chua Yong Tong Keyboard error reduction method and apparatus
US7665041B2 (en) 2003-03-25 2010-02-16 Microsoft Corporation Architecture for controlling a computer using hand gestures
DE602004006190T8 (en) 2003-03-31 2008-04-10 Honda Motor Co., Ltd. Device, method and program for gesture recognition
EP1626877A4 (en) 2003-03-31 2011-08-10 Timothy R Pryor Reconfigurable vehicle instrument panels
CA2426867A1 (en) 2003-04-30 2004-10-30 Naviform Holdings Ltd. Customizable keyboard
US7884804B2 (en) 2003-04-30 2011-02-08 Microsoft Corporation Keyboard with input-sensitive display device
US7233316B2 (en) 2003-05-01 2007-06-19 Thomson Licensing Multimedia user interface
US8555165B2 (en) 2003-05-08 2013-10-08 Hillcrest Laboratories, Inc. Methods and systems for generating a zoomable graphical user interface
US8479112B2 (en) 2003-05-13 2013-07-02 Microsoft Corporation Multiple input language selection
US7302642B2 (en) 2003-06-03 2007-11-27 Tim Simon, Inc. Thermostat with touch-screen display
US7310781B2 (en) 2003-06-05 2007-12-18 International Business Machines Corporation System and method for content and information transfer between program entities
US7475390B2 (en) 2004-01-12 2009-01-06 International Business Machines Corporation System and method for automatic natural language translation during information transfer
US20060242607A1 (en) 2003-06-13 2006-10-26 University Of Lancaster User interface
JP3938104B2 (en) 2003-06-19 2007-06-27 ヤマハ株式会社 Arpeggio pattern setting device and program
US20040263484A1 (en) 2003-06-25 2004-12-30 Tapio Mantysalo Multifunctional UI input device for moblie terminals
US7474772B2 (en) 2003-06-25 2009-01-06 Atrua Technologies, Inc. System and method for a miniature user input device
US7057607B2 (en) 2003-06-30 2006-06-06 Motorola, Inc. Application-independent text entry for touch-sensitive display
US8373660B2 (en) 2003-07-14 2013-02-12 Matt Pallakoff System and method for a portable multimedia client
US20050015731A1 (en) 2003-07-15 2005-01-20 Microsoft Corporation Handling data across different portions or regions of a desktop
US7499040B2 (en) 2003-08-18 2009-03-03 Apple Inc. Movable touch pad with added functionality
US20060181517A1 (en) 2005-02-11 2006-08-17 Apple Computer, Inc. Display actuator
US9024884B2 (en) 2003-09-02 2015-05-05 Apple Inc. Touch-sensitive electronic apparatus for media applications, and methods therefor
US20050052427A1 (en) 2003-09-10 2005-03-10 Wu Michael Chi Hung Hand gesture interaction with touch surface
US7411575B2 (en) 2003-09-16 2008-08-12 Smart Technologies Ulc Gesture recognition method and touch system incorporating the same
JP2005108041A (en) 2003-09-30 2005-04-21 Toshiba Corp Method for displaying menu screen on portable terminal and portable terminal
US7209116B2 (en) 2003-10-08 2007-04-24 Universal Electronics Inc. Control device having integrated mouse and remote control capabilities
US20050091577A1 (en) 2003-10-23 2005-04-28 International Business Machines Corporation Information integration system
US7478336B2 (en) 2003-11-06 2009-01-13 International Business Machines Corporation Intermediate viewer for transferring information elements via a transfer buffer to a plurality of sets of destinations
US6970749B1 (en) 2003-11-12 2005-11-29 Adobe Systems Incorporated Grouped palette stashing
US7495659B2 (en) 2003-11-25 2009-02-24 Apple Inc. Touch pad for handheld device
JP2007517291A (en) 2003-12-31 2007-06-28 リサーチ イン モーション リミテッド Keyboard layout
US7346853B2 (en) 2004-01-12 2008-03-18 International Business Machines Corporation Online learning monitor
US7340685B2 (en) 2004-01-12 2008-03-04 International Business Machines Corporation Automatic reference note generator
US20070159453A1 (en) 2004-01-15 2007-07-12 Mikio Inoue Mobile communication terminal
US20050162402A1 (en) 2004-01-27 2005-07-28 Watanachote Susornpol J. Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback
US7697729B2 (en) 2004-01-29 2010-04-13 Authentec, Inc. System for and method of finger initiated actions
US6856259B1 (en) 2004-02-06 2005-02-15 Elo Touchsystems, Inc. Touch sensor system to detect multiple touch events
US7545784B2 (en) 2004-02-11 2009-06-09 Yahoo! Inc. System and method for wireless communication between previously known and unknown users
EP2254026A1 (en) 2004-02-27 2010-11-24 Research In Motion Limited Text input system for a mobile electronic device and methods thereof
US7706616B2 (en) 2004-02-27 2010-04-27 International Business Machines Corporation System and method for recognizing word patterns in a very large vocabulary based on a virtual keyboard layout
US7607576B2 (en) 2004-02-27 2009-10-27 Gilbarco, Inc. Local zone security architecture for retail environments
US20050190970A1 (en) 2004-02-27 2005-09-01 Research In Motion Limited Text input system for a mobile electronic device and methods thereof
US20050204008A1 (en) 2004-03-09 2005-09-15 Marc Shinbrood System and method for controlling the downstream preservation and destruction of electronic mail
US7183476B2 (en) 2004-03-18 2007-02-27 Swingle Margaret J Portable electronic music score device for transporting, storing displaying, and annotating music scores
US7180502B2 (en) 2004-03-23 2007-02-20 Fujitsu Limited Handheld device with preferred motion selection
CN100565433C (en) 2004-05-05 2009-12-02 皇家飞利浦电子股份有限公司 Browsing media items
US7519223B2 (en) 2004-06-28 2009-04-14 Microsoft Corporation Recognizing gestures and using gestures for interacting with software applications
WO2006003590A2 (en) 2004-06-29 2006-01-12 Koninklijke Philips Electronics, N.V. A method and device for preventing staining of a display device
US7743348B2 (en) 2004-06-30 2010-06-22 Microsoft Corporation Using physical objects to adjust attributes of an interactive display application
US20060007174A1 (en) 2004-07-06 2006-01-12 Chung-Yi Shen Touch control method for a drag gesture and control module thereof
US7398472B2 (en) 2004-07-09 2008-07-08 Microsoft Corporation Defining the visual appearance of user-interface controls
US7664748B2 (en) 2004-07-12 2010-02-16 John Eric Harrity Systems and methods for changing symbol sequences in documents
US20060012577A1 (en) 2004-07-16 2006-01-19 Nokia Corporation Active keypad lock for devices equipped with touch screen
US20080129707A1 (en) 2004-07-27 2008-06-05 Pryor Timothy R Method and apparatus employing multi-functional controls and displays
AU2005269406A1 (en) 2004-07-28 2006-02-09 Newton Peripherals, Llc Peripheral devices for portable computer
US7653883B2 (en) 2004-07-30 2010-01-26 Apple Inc. Proximity detector in handheld device
KR100958490B1 (en) 2004-07-30 2010-05-17 애플 인크. Mode-based graphical user interfaces for touch sensitive input devices
KR101270847B1 (en) 2004-07-30 2013-06-05 애플 인크. Gestures for touch sensitive input devices
US8381135B2 (en) 2004-07-30 2013-02-19 Apple Inc. Proximity detector in handheld device
US7599044B2 (en) 2005-06-23 2009-10-06 Apple Inc. Method and apparatus for remotely detecting presence
US7620705B2 (en) * 2004-07-30 2009-11-17 Research In Motion Limited Method and apparatus for provisioning a communications client on a host device
KR100737558B1 (en) 2004-08-16 2007-07-10 주식회사 팬택앤큐리텔 Wireless telecommunication terminal and method for controlling alarm volume using preference level
KR101065943B1 (en) 2004-08-16 2011-09-20 애플 인크. A method of increasing the spatial resolution of touch sensitive devices
US20060044280A1 (en) 2004-08-31 2006-03-02 Huddleston Wyatt A Interface
US20100231506A1 (en) 2004-09-07 2010-09-16 Timothy Pryor Control of appliances, kitchen and home
US7719522B2 (en) 2004-09-24 2010-05-18 Apple Inc. Raw data track pad device and system
US20060066590A1 (en) 2004-09-29 2006-03-30 Masanori Ozawa Input device
US20060071915A1 (en) 2004-10-05 2006-04-06 Rehm Peter H Portable computer and method for taking notes with sketches and typed text
US7454717B2 (en) 2004-10-20 2008-11-18 Microsoft Corporation Delimiters for selection-action pen gesture phrases
US7242311B2 (en) 2004-10-29 2007-07-10 Caterpillar Inc. Method and system for providing work machine multi-functional user interface
US7489306B2 (en) 2004-12-22 2009-02-10 Microsoft Corporation Touch screen accuracy
US7617283B2 (en) 2005-01-10 2009-11-10 International Business Machines Corporation System and method for instant messaging
US20060181519A1 (en) 2005-02-14 2006-08-17 Vernier Frederic D Method and system for manipulating graphical objects displayed on a touch-sensitive display surface using displaced pop-ups
US7605804B2 (en) 2005-04-29 2009-10-20 Microsoft Corporation System and method for fine cursor positioning using a low resolution imaging touch screen
US7487461B2 (en) 2005-05-04 2009-02-03 International Business Machines Corporation System and method for issuing commands based on pen motions on a graphical keyboard
US8185841B2 (en) 2005-05-23 2012-05-22 Nokia Corporation Electronic text input involving a virtual keyboard and word completion functionality on a touch-sensitive display screen
US20070011603A1 (en) 2005-07-06 2007-01-11 Mikko Makela Method, system, device and software product for showing tooltips for page segments and generating content for the page segments
WO2007016370A2 (en) 2005-07-29 2007-02-08 Eazypaper Inc. Computer method and apparatus using embedded message window for displaying messages in a functional bar
US7737999B2 (en) 2005-08-26 2010-06-15 Veveo, Inc. User interface for visual cooperation between text input and display device
US7443316B2 (en) 2005-09-01 2008-10-28 Motorola, Inc. Entering a character into an electronic device
US7694231B2 (en) 2006-01-05 2010-04-06 Apple Inc. Keyboards for portable electronic devices
US8745140B2 (en) 2005-10-14 2014-06-03 Blackberry Limited System and method of handling messages for forwarding
US8819144B2 (en) 2005-10-14 2014-08-26 Blackberry Limited Method, device, software and graphical user interface for forwarding messages between message handling services
US7958456B2 (en) 2005-12-23 2011-06-07 Apple Inc. Scrolling list with floating adjacent index symbols
US8139028B2 (en) 2006-02-01 2012-03-20 Synaptics Incorporated Proximity sensor and method for indicating extended interface results
EP1816573A1 (en) 2006-02-02 2007-08-08 Nextair Corporation Apparatus, method and machine-readable medium for facilitating generation of a markup language document containing identical sets of markup language elements
US7529795B2 (en) 2006-03-20 2009-05-05 Stragent, Llc Message board aggregator
US7538760B2 (en) 2006-03-30 2009-05-26 Apple Inc. Force imaging input device and system
US7511702B2 (en) 2006-03-30 2009-03-31 Apple Inc. Force and location sensitive display
US7978181B2 (en) 2006-04-25 2011-07-12 Apple Inc. Keystroke tactility arrangement on a smooth touch surface
US8279180B2 (en) 2006-05-02 2012-10-02 Apple Inc. Multipoint touch surface controller
US7765494B2 (en) 2006-05-24 2010-07-27 Sap Ag Harmonized theme definition language
US20080040692A1 (en) 2006-06-29 2008-02-14 Microsoft Corporation Gesture input
US8694895B2 (en) 2007-02-05 2014-04-08 Microsoft Corporation Human interaction with application from email client
US8209619B2 (en) 2007-03-16 2012-06-26 Oracle International Corporation Non-linear navigation in discussion forums
US9740386B2 (en) 2007-06-13 2017-08-22 Apple Inc. Speed/positional mode translations
US8059101B2 (en) 2007-06-22 2011-11-15 Apple Inc. Swipe gestures for touch screen keyboards
US20090064006A1 (en) 2007-08-30 2009-03-05 Indran Naick Techniques for Performing Tasks Associated with Custom Folder Controls
JP5533254B2 (en) 2010-05-24 2014-06-25 アイシン・エィ・ダブリュ株式会社 Information display device, information display method, and program
US8773377B2 (en) 2011-03-04 2014-07-08 Microsoft Corporation Multi-pass touch contact tracking
WO2014043275A1 (en) 2012-09-11 2014-03-20 Apple Inc. Gesturing with a multipoint sensing device

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI474226B (en) * 2008-12-17 2015-02-21 Htc Corp Portable communication device and method for adjusting a plurality of touch signals thereof
TWI395124B (en) * 2009-04-02 2013-05-01 Mstar Semiconductor Inc Digitizing apparatus, digital converting method and capacitive touch panel apparatus
TWI469019B (en) * 2009-04-30 2015-01-11 Chimei Innolux Corp Power saving apparatus for capacitive touch screen and power saving method for same and capacitive touch screen employing power saving apparatus
US9377890B2 (en) 2009-05-11 2016-06-28 Au Optronics Corp. Multi-touch method for resistive touch panel
CN101943990A (en) * 2009-07-03 2011-01-12 索尼公司 Operating control device, method of controlling operation thereof and computer program
TWI494791B (en) * 2009-11-06 2015-08-01 Au Optronics Corp Method of determining gestures for touch device
TWI493407B (en) * 2009-11-09 2015-07-21 Elan Microelectronics Corp Multi - function touchpad remote control and its control method
TWI413922B (en) * 2010-04-23 2013-11-01 Primax Electronics Ltd Control method for touchpad and touch device using the same
US9134897B2 (en) 2010-04-26 2015-09-15 Via Technologies, Inc. Electronic system and method for operating touch screen thereof
TWI470537B (en) * 2010-12-20 2015-01-21 Apple Inc Event recognition method, related electronic device and computer readable storage medium
TWI455002B (en) * 2011-02-01 2014-10-01 Edamak Corp Device and method for proximity gesture detection
TWI470532B (en) * 2011-06-09 2015-01-21 Inventec Corp Electronic device and control method thereof
TWI488487B (en) * 2011-10-18 2015-06-11 Acer Inc Method for adjusting video compression using gesture
CN104106030A (en) * 2011-12-22 2014-10-15 纳米技术方案公司 Switched-electrode capacitive-measurement device for touch-sensitive and contactless interfaces
TWI480797B (en) * 2013-08-14 2015-04-11 Pegatron Corp Touch method and electronic apparatus
US10402144B2 (en) 2017-05-16 2019-09-03 Wistron Corporation Portable electronic device and operation method thereof

Also Published As

Publication number Publication date
WO2008094791A9 (en) 2008-09-18
WO2008094791A3 (en) 2008-11-27
EP2115560A2 (en) 2009-11-11
AU2008210864A1 (en) 2008-08-07
TWM350737U (en) 2009-02-11
CN101636711A (en) 2010-01-27
US20080036743A1 (en) 2008-02-14
JP2010517197A (en) 2010-05-20
HK1109023A2 (en) 2008-05-30
DE202008001338U1 (en) 2008-07-10
US9292111B2 (en) 2016-03-22
WO2008094791A2 (en) 2008-08-07
AU2008100085A4 (en) 2008-03-06

Similar Documents

Publication Publication Date Title
TW200847001A (en) Gesturing with a multipoint sensing device
KR101072762B1 (en) Gesturing with a multipoint sensing device
US9239673B2 (en) Gesturing with a multipoint sensing device
US9348458B2 (en) Gestures for touch sensitive input devices
US8842084B2 (en) Gesture-based object manipulation methods and devices
US9104308B2 (en) Multi-touch finger registration and its applications
WO2014043275A1 (en) Gesturing with a multipoint sensing device
Uddin Improving Multi-Touch Interactions Using Hands as Landmarks
US10133346B2 (en) Gaze based prediction device and method
AU2016238971B2 (en) Gesturing with a multipoint sensing device
AU2014201419B2 (en) Gesturing with a multipoint sensing device