US20140010042A1 - Sonar Preview Mode - Google Patents
Sonar Preview Mode Download PDFInfo
- Publication number
- US20140010042A1 US20140010042A1 US13/759,760 US201313759760A US2014010042A1 US 20140010042 A1 US20140010042 A1 US 20140010042A1 US 201313759760 A US201313759760 A US 201313759760A US 2014010042 A1 US2014010042 A1 US 2014010042A1
- Authority
- US
- United States
- Prior art keywords
- sonar images
- pane
- sonar
- images
- recorded
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims abstract description 31
- 238000005516 engineering process Methods 0.000 abstract description 17
- 238000010586 diagram Methods 0.000 description 17
- 230000003287 optical effect Effects 0.000 description 8
- 238000004891 communication Methods 0.000 description 6
- 238000012545 processing Methods 0.000 description 6
- 238000012552 review Methods 0.000 description 6
- 238000004091 panning Methods 0.000 description 5
- 230000004044 response Effects 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 230000003993 interaction Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 3
- 230000005055 memory storage Effects 0.000 description 2
- 230000006855 networking Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000010079 rubber tapping Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007723 transport mechanism Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B42/00—Obtaining records using waves other than optical waves; Visualisation of such records by using optical means
- G03B42/06—Obtaining records using waves other than optical waves; Visualisation of such records by using optical means using ultrasonic, sonic or infrasonic waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/56—Display arrangements
- G01S7/60—Display arrangements for providing a permanent recording
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/56—Display arrangements
- G01S7/62—Cathode-ray tube displays
- G01S7/6218—Cathode-ray tube displays providing two-dimensional coordinated display of distance and direction
- G01S7/6236—Sector-scan displays
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/56—Display arrangements
- G01S7/62—Cathode-ray tube displays
- G01S7/6272—Cathode-ray tube displays producing cursor lines and indicia by electronic means
Definitions
- a log of sonar images can be very useful for marine activities.
- the operator of a marine vessel monitors marine traffic.
- a device that is easy to operate can provide advantages to a vessel operator.
- a device that provides data in an easy to follow format can also provide advantages to the vessel operator. Such advantages may include requiring less time to request and review information, which in turn provides the vessel operator with more time to monitor the marine traffic.
- sonar images that are currently being recorded may be displayed on a regular pane. Further, condensed sonar images that had previously been recorded and that are currently being recorded may be displayed on a preview pane, while recording the sonar images.
- the apparatus includes one or more processors, a screen, and a memory.
- the memory has a plurality of executable instructions.
- the processors may display sonar images that are being recorded at a present time, receive a request to display past sonar images that had been recorded at a previous time, and display the past sonar images while still causing the sonar images that are currently being captured at the present time to be recorded.
- Described herein are also implementations of various technologies for a non-transitory computer readable medium having stored thereon computer-executable instructions which, when executed by a computer, cause the computer to: display sonar images that are currently being recorded on a regular pane; and display condensed sonar images that had previously been recorded and that are currently being recorded on a preview pane, while recording the sonar images.
- a computing system includes at least one processor, at least one memory, and one or more programs stored in the at least one memory; and means for performing any method disclosed herein.
- FIG. 1 is a schematic diagram of a display device in accordance with various techniques described therein.
- FIG. 2 is a timing diagram in accordance with various techniques described therein.
- FIG. 3 is a schematic diagram of a graphical user interface showing past sonar images in accordance with various techniques described therein.
- FIG. 4 is a flow diagram describing the display of past sonar images in accordance with various techniques described therein.
- FIG. 5 is a schematic diagram of a graphical user interface with a preview window in accordance with various techniques described therein.
- FIG. 6 is a schematic diagram of a graphical user interface showing past sonar images using a preview window in accordance with various techniques described therein.
- FIG. 7 is a flow diagram for showing past sonar images using a preview window in accordance with various techniques described therein.
- FIG. 8 is a flow diagram describing the preview window modes in accordance with various techniques described therein.
- FIG. 9 is a schematic diagram of a computing system in accordance with various techniques described therein.
- first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another.
- a first object or step could be termed a second object or step, and, similarly, a second object or step could be termed a first object or step, without departing from the scope of the invention.
- the first object or step, and the second object or step are both objects or steps, respectively, but they are not to be considered the same object or step.
- the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in response to detecting,” depending on the context.
- the phrase “if it is determined” or “if [a stated condition or event] is detected” may be construed to mean “upon determining” or “in response to determining” or “upon detecting [the stated condition or event]” or “in response to detecting [the stated condition or event],” depending on the context.
- various technologies and techniques described herein include receiving user requests for a number of different operations.
- the user request for a particular operation will be explicitly described.
- “user request” or “user can request” shall also include, but are not limited to, touching the screen, double tapping the screen (tapping the screen twice in rapid succession), pressing a particular physical or virtual button, making a selection from a menu, placement of a cursor at a particular location, stylus pointing, mouse selection, an audible command, as well as the explicit description of the “user request” for the particular operation or how the “user can request”, explicitly, the particular operations.
- various technologies and techniques described herein are directed to a display device having a screen.
- the screen can display sonar images that are currently being captured as well as sonar images that had previously been captured.
- the sonar images form a stream of sonar images that may be ordered based on time.
- the display device may provide the user access to older portions of sonar images that were captured considerably long before the currently captured sonar images. For example, the user may access to these older portions by placing their finger on the screen over the sonar data and swiping in the direction of the most recent sonar image. The sonar images on the screen will move according to the movement of the finger. Sonar images on the right (the most recent) will move off the screen, and the older sonar images will appear on the left hand side of the screen. The user can continue this process (swiping the screen and moving the displayed sonar data) until the sonar images from a desired time appear on the screen.
- the display device may continue to record captured sonar images and add them to the sonar image stream.
- the user can return to viewing the sonar images that are currently being captured.
- the user can view the sonar images that are currently being captured while the user is viewing older sonar images by swiping the displayed sonar images to the left.
- the oldest sonar images drop off the right hand side and newer sonar images appear on the left hand side.
- the sonar image stream can be quite lengthy and the user may have to search a long time to view a particular sonar image of interest.
- the display device may provide a preview of the entire sonar image stream in a smaller region of the screen, e.g., a preview pane.
- the sonar images in the preview pane may be condensed.
- the condensed sonar images may be large enough for the user to notice some of the significant characteristics.
- the user can identify the desired sonar image by spotting the significant characteristics in the condensed sonar image stream. To select the desired sonar image for view in more detail in a regular pane, the user can highlight the desired area in the condensed sonar image stream, and the corresponding sonar image in regular size will appear in the regular pane.
- FIG. 1 illustrates a block diagram of a display device 100 in accordance with various implementations described herein.
- the display device includes a screen 105 that provides a graphical user interface for displaying sonar images 110 .
- Sonar images 110 may provide a map of the underlying body of water in the vicinity of a sonar device.
- the display device 100 may also include a plurality of buttons 112 to facilitate user interaction.
- the plurality of buttons 112 can be physical buttons (shown) or virtual buttons (not shown).
- the display device 100 can include a speaker/microphone 113 .
- the screen 105 can be touch-sensitive or sensitive to the human body heat of a finger.
- the display device 100 may be able to detect the location of the screen 105 that is touched by the finger (not shown).
- the display device 100 may provide a cursor 114 .
- the cursor 114 may be movable by placement of the finger over the cursor 114 and dragging the finger to a desired location. The display device 100 may cause the cursor 114 to follow the finger.
- the sonar images 110 may be presented as a two-dimensional graphics, where the vertical axis 115 represents depth and the horizontal axis 120 represents time.
- the graphical user interface may provide a number of tools to allow the user to obtain more specific information. For example, the user can use the cursor 114 for pointing to a particular location in the graphical user interface. In certain implementations, the display device 100 presents the cursor 114 in response to receiving a user request for the cursor 114 .
- the graphical user interface can provide specific information about the particular location, such as depth 130 and estimated water temperature 132 at the particular location.
- the graphical user interface can also provide ocean depth 155 , estimated water temperature 160 , latitude 170 and longitude 175 at a current location of a vessel carrying the display device 100 .
- the graphical user interface can also provide zooming 180 for more detail or more range.
- Sonar equipment can capture sonar images in real time.
- the display device 100 is configured to receive the sonar images and record the sonar images to a recording medium (not shown), such as memory, and also provide the sonar images for display on the screen 105 in real time.
- the newly captured sonar images 110 appear on the rightmost part of the graphic, while the older sonar images 110 appear towards the left. After some time, the older sonar images 110 on the left move off of the screen.
- the right side has more recent sonar images 110 while the left side has older sonar images 110 .
- the foregoing can be reversed.
- the recording medium can be part of the display device, while in other implementations, the recording medium can be external to the display device 100 .
- the display device 100 allows the user to access a previously captured and recorded sonar images.
- FIG. 2 illustrates a timing diagram showing different times that sonar images are recorded and displayed.
- Axis 200 indicates time, t 0 . . . t n , wherein t 0 comes first, t 1 comes second, t 2 comes third, etc.
- the sonar images 110 received by the sonar device can be associated with time. For example, at time t 4 , the user may request the display device 100 to display the sonar images collected at time t 0 to t 1 , which were collected earlier than time t 4 .
- the user may review the sonar images collected between time t 0 to t 1 from t 4 to t 8 While the user reviews the sonar images at t 0 to t 1 , the sonar images from t 4 to t 8 are continually being captured and recorded. These sonar images are retrievable for viewing by the user at a later time. For example, at time t 8 , the user can view sonar images captured at time t6.
- the display device 100 may provide the user with at least two ways of requesting access to past sonar images 110 . These include a panning mode and a preview mode.
- FIG. 3 illustrates a schematic diagram of a graphical user interface showing past sonar images 110 by panning.
- the user places their finger 305 onto the screen 105 over the sonar images 110 and swipes their finger 305 in a particular direction 310 .
- the sonar images 110 move in the direction of the finger 305 . This is known as panning. Certain portions of the sonar images 110 will no longer appear on the screen, while new portions of the recorded sonar images will appear on the screen 105 .
- the data will not move to the left if the most recent sonar image 110 is displayed. Additionally, the sonar images 110 will not move to the right once the user reaches the oldest sonar images 110 that are stored.
- the user can use one or more of the plurality of buttons 112 . For example, some of the plurality of buttons 112 can be associated with directions.
- FIG. 4 is a flow diagram for reviewing past sonar images 110 in accordance with various implementations described herein. The flow diagram will be described with references to FIGS. 1-3 .
- the display device 100 displays current sonar images 110 .
- the display device 100 receives a user request to move the sonar images 110 backwards, or to view past sonar images 110 .
- the user request may include placement of the finger 305 on the screen 105 and swiping the finger to the right.
- the user may wish to review the sonar images 110 captured at t 0 .
- the desired past sonar images 110 captured at t 0 may be displayed on the screen 105 .
- the display device 100 displays the past sonar images 110 in accordance with the user request, while also recording current sonar images.
- the display device 100 can display the past sonar images 110 by receiving finger swipes that cause the display device to continuously pan the sonar images 110 .
- the display device 100 receives a user request to review sonar images 110 that were just recorded while the user was viewing past sonar images 110 during blocks 410 - 415 .
- the display device 100 was recording sonar images captured from times t 4 -t 8 .
- the user may wish to review the sonar images captured at time t 6 .
- the user request can include swiping the finger to the left. The user may continue swiping/panning until the display device 100 displays sonar images captured at time t 6 .
- the display device 100 displays the sonar images 110 requested by the user during block 420 , e.g., sonar images captured at time t 6 .
- the display device 100 may receive a user request to display the most current sonar images. This user request may be made for example at time t 9 .
- the user request can include swiping the screen to the left until the most recent sonar images 110 , i.e., sonar images captured at time t 9 , are displayed on the display device 100 .
- the display device 100 displays the most current sonar images, e.g., t 9 . . . t n .
- FIG. 5 is a schematic of a graphical user interface with a preview pane 505 in accordance with various implementations described herein.
- the preview pane 505 displays previously recorded sonar images 110 at a size that is smaller in size to the sonar images 110 displayed in a regular pane 515 .
- the preview pane 505 can include sonar images 110 covering a longer period of time than the regular pane 515 .
- the preview pane 505 can include a condensed version of all the sonar images 110 (condensed sonar images) that have been recorded.
- the recording medium may have a capacity that limits the time period that the sonar images 110 can be stored. For example, the time period can be about ten minutes.
- the preview pane 505 can provide condenses sonar images 110 from the most recent hour. Although the preview pane 505 is described herein as having condensed sonar images, it should be understood that the sonar images may also be in compressed format.
- the preview pane 505 may include a semi-transparent sliding box 510 that identifies the portion of the preview pane 505 representing the sonar images 110 that are shown in the regular pane 515 .
- the sliding box 510 is all the way to the right.
- FIG. 6 illustrates viewing past sonar images 110 using the preview pane 505 in accordance with various implementations described herein.
- the condensed sonar images in the preview pane 505 are smaller than the sonar images 110 in the regular pane 515 , but are still large enough for the user to scan and locate notable features.
- the condensed sonar images in the sliding box 510 are displayed at regular size in the regular pane 515 .
- moving the sliding box 510 moves the sonar images 110 in the regular pane 515 to correspond to the condensed sonar images in the sliding box 510 .
- the user can switch the positions of the sliding box 510 by placing a finger 305 on the desired condensed sonar images in the preview pane 505 .
- the sliding box 510 is displayed around that area and the sonar images in regular size that correspond to the desired condensed sonar images are displayed at the regular pane 515 .
- the sonar images 110 in the regular pane 515 freeze until the user releases the sliding box 510 .
- the sonar images displayed in the regular pane 515 correspond to the condensed sonar images in the sliding box 510 .
- FIG. 7 is a flow diagram for reviewing past sonar images 110 using the preview pane 505 in accordance with various implementations described herein.
- the display device 100 displays sonar images 110 that are currently captured in a regular pane 515 .
- the display device 100 displays condensed sonar images that were previously recorded in the preview pane 505 , while displaying currently recorded sonar images 110 .
- the display device 100 highlights a first portion of the condensed sonar images on the preview pane 505 with a semi-transparent box, such as sliding box 510 .
- the display device 100 displays the sonar images that correspond to the first portion on the regular pane 515 in regular format.
- the display device 100 may receive a user request to highlight a second portion of the condensed sonar images on the preview pane 505 .
- the user request can include receiving finger contact with the sliding box 510 and swiping the sliding box 510 .
- the user request can include receiving direct finger contact at the second portion of the preview pane 505 .
- the display device 100 highlights the second portion of the condensed sonar images on the preview pane 505 .
- the display device 100 displays the sonar images in the regular pane 515 that correspond to the condensed sonar images in the second portion of the preview pane 505 .
- the graphical user interface can include a menu 605 with a “Back” selection 610 ( 1 ), an “Off” selection 610 ( 2 ), a “Cursor Only” selection 610 ( 3 ), and an “Always” selection 610 ( 4 ).
- the display device 100 can be multifunctional and provide additional functions to just sonar image presentation.
- the “Back” selection 610 ( 1 ) allows the user to leave the sonar image presentation function and access a higher level menu.
- the display device 100 can provide the preview pane 505 , based on user selections. For instance, if the user selects the “Off” selection 610 ( 2 ), the display device 100 does not display the preview pane 505 . The user can also request that the display device 100 always display the preview pane 505 by selection the “Always” selection 610 ( 4 ). In some implementations, the preview pane 505 appears based on whether there is a cursor 114 . If there is no cursor 114 , the preview pane 505 does not appear. If there is a cursor 114 present, the preview pane 505 appears. The user can request display of the preview pane based on existence of the cursor 114 by selecting the “Cursor Only” selection 610 ( 3 ).
- FIG. 8 is a flow diagram for determining whether to place the preview pane 505 or not.
- the display device 100 determines whether the mode is “Off”, “Always”, or “Cursor Only”. If at block 805 , the display device 100 determines that the mode is “Off”, at block 810 , the display device 100 does not display the preview pane 505 and returns to block 805 . If at block 805 , the display device 100 determines that the mode is “Always”, at block 815 , the display device 100 displays the preview pane 505 and returns to block 805 .
- the display device 100 determines that the display device 100 is in “Cursor Only” mode, the display device 100 makes a determination whether the cursor is present (block 820 ). If the cursor 114 is present, the display device 100 displays the preview pane 505 at block 815 and returns to block 805 . If the cursor is not present at block 820 , the display device 100 does not display the preview pane 505 (block 815 ). The display device 100 then returns to block 805 .
- Implementations of various technologies described herein may be operational with numerous general purpose or special purpose computing system environments or configurations.
- Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the various technologies described herein include, but are not limited to, personal computers, server computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
- program modules include routines, programs, objects, components, data structures, etc. that performs particular tasks or implement particular abstract data types. Further, each program module may be implemented in its own way, and all need not be implemented the same way. While program modules may all execute on a single computing system, it should be appreciated that, in some implementations, program modules may be implemented on separate computing systems or devices adapted to communicate with one another. A program module may also be some combination of hardware and software where particular tasks performed by the program module may be done either through hardware, software, or both.
- program modules may be located in both local and remote computer storage media including memory storage devices.
- FIG. 9 illustrates a schematic diagram of a computing system 900 in which the various technologies described herein may be incorporated and practiced.
- the computing system 900 may be a conventional desktop, a handheld device, personal digital assistant, a server computer, electronic device/instrument, laptop, tablet, or part of a navigation system, marine electronics, or sonar system. It should be noted, however, that other computer system configurations may be used.
- the computing system 900 may include a central processing unit (CPU) 930 , a system memory 926 , a graphics processing unit (GPU) 931 and a system bus 928 that couples various system components including the system memory 926 to the CPU 930 . Although only one CPU is illustrated in FIG. 9 , it should be understood that in some implementations the computing system 900 may include more than one CPU.
- CPU central processing unit
- GPU graphics processing unit
- the CPU 930 can include a microprocessor, a microcontroller, a processor, a programmable integrated circuit, or a combination thereof.
- the CPU 930 can comprise an off-the-shelf processor such as a Reduced Instruction Set Computer (RISCTM), or a Microprocessor without Interlocked Pipeline Stages (MIPSTM) processor, or a combination thereof.
- RISCTM Reduced Instruction Set Computer
- MIPSTM Microprocessor without Interlocked Pipeline Stages
- the CPU 930 can also comprise a proprietary processor.
- the GPU 931 may be a microprocessor specifically designed to manipulate and implement computer graphics.
- the CPU 930 may offload work to the GPU 931 .
- the GPU 931 may have its own graphics memory, and/or may have access to a portion of the system memory 926 .
- the GPU 931 may include one or more processing units, and each processing unit may include one or more cores.
- the CPU 930 can provide output data to a graphics processor 931 .
- the graphics processor 931 can generate graphical user interfaces that present the output data.
- the graphics processor 931 can also provide objects, such as menus, in the graphical user interface. A user can provide inputs by interacting with the objects.
- the graphics processor 1531 can receive the inputs from interaction with the objects and provide the inputs to the CPU 930 .
- the system bus 928 may be any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
- bus architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus also known as Mezzanine bus.
- the system memory 926 may include a read only memory (ROM) 912 and a random access memory (RAM) 916 .
- a basic input/output system (BIOS) 914 containing the basic routines that help transfer information between elements within the computing system 900 , such as during start-up, may be stored in the ROM 912 .
- the computing system 900 may further include a hard disk drive interface 936 for reading from and writing to a hard disk 950 , a magnetic disk drive 952 for reading from and writing to a removable magnetic disk 956 , and an optical disk drive 954 for reading from and writing to a removable optical disk 958 , such as a CD ROM or other optical media.
- the hard disk 950 , the magnetic disk drive 952 , and the optical disk drive 954 may be connected to the system bus 928 by a hard disk drive interface 936 , a magnetic disk drive interface 938 , and an optical drive interface 940 , respectively.
- the drives and their associated computer-readable media may provide nonvolatile storage of computer-readable instructions, data structures, program modules and other data for the computing system 900 .
- computing system 900 may also include other types of computer-readable media that may be accessed by a computer.
- computer-readable media may include computer storage media and communication media.
- Computer storage media may include volatile and non-volatile, and removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules or other data.
- Computer storage media may further include RAM, ROM, erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other solid state memory technology, CD-ROM, digital versatile disks (DVD), or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computing system 900 .
- Communication media may embody computer readable instructions, data structures, program modules or other data in a modulated data signal, such as a carrier wave or other transport mechanism and may include any information delivery media.
- modulated data signal may mean a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.
- the computing system 900 may also include a host adapter 933 that connects to a storage device 935 via a small computer system interface (SCSI) bus, a Fiber Channel bus, an eSATA bus, or using any other applicable computer bus interface. Combinations of any of the above may also be included within the scope of computer readable media.
- SCSI small computer system interface
- a number of program modules may be stored on the hard disk 950 , magnetic disk 956 , optical disk 958 , ROM 912 or RAM 916 , including an operating system 918 , one or more application programs 920 , control pilot 922 , and program data 924 .
- the hard disk 950 may include and a database system.
- the database system could record, for example, recorded points.
- the application programs 920 may include various mobile applications (“apps”) and other applications configured to perform various methods and techniques described herein.
- the operating system 918 may be any suitable operating system that may control the operation of a networked personal or server computer.
- a user may enter commands and information into the computing system 900 through input devices such as buttons 962 .
- Other input devices may include a microphone (not shown).
- These and other input devices may be connected to the CPU 930 through a serial port interface 942 coupled to system bus 928 , but may be connected by other interfaces, such as a parallel port, game port or a universal serial bus (USB).
- USB universal serial bus
- the user can enter inputs through a monitor 934 .
- the screen 905 can be sensitive to touching by a finger 110 , while in other implementations, the screen 905 can be sensitive to the body heat from the finger 110 .
- the screen 905 detects placement and movement of the user's fingers and provides signals that are indicative of such placement and movement to the CPU 930 . Additionally, the screen 905 detects user interactions with the objects provided by the GPU 931 and provides signals that are indicative of the user interactions with the objects.
- the monitor 934 may also be connected to system bus 928 via an interface, such as a video adapter 932 .
- the computing system 900 may operate in a networked environment using logical connections to one or more remote computers 974 .
- the logical connections may be any connection that is commonplace in offices, enterprise-wide computer networks, intranets, and the Internet, such as local area network (LAN) 976 and a wide area network (WAN) 966 .
- the remote computers 974 may be another computer, a server computer, a router, a network PC, a peer device or other common network node, and may include many or all of the elements describes above relative to the computing system 900 .
- the remote computers 974 may also each include application programs 970 similar to that of the computer action function.
- the computing system 900 may be connected to the local network 976 through a network interface or adapter 944 .
- the computing system 900 may include a wireless router 964 or other means for establishing communication over a wide area network 966 , such as the Internet.
- the router 964 may be connected to the system bus 928 via the serial port interface 942 .
- program modules depicted relative to the computing system 900 may be stored in a remote memory storage device 972 . It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers may be used.
- the computing system 900 can receive sonar signals from a sonar system through the network interface 944 .
- the central processing unit 930 , GPU 931 , video adapter 932 , network interface 944 , and system memory 926 , hard disc 950 , monitor 934 , and screen 905 can be integrated into a console.
Landscapes
- Engineering & Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Acoustics & Sound (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- This application claims the benefit of U.S. Provisional Patent Application Ser. No. 61/668,965, filed Jul. 6, 2012, titled SONAR PREVIEW MODE, and the disclosure of which is incorporated herein by reference.
- This section is intended to provide background information to facilitate a better understanding of various technologies described herein. As the section's title implies, this is a discussion of related art. That such art is related in no way implies that it is prior art. The related art may or may not be prior art. It should therefore be understood that the statements in this section are to be read in this light, and not as admissions of prior art.
- A log of sonar images can be very useful for marine activities. The operator of a marine vessel monitors marine traffic. A device that is easy to operate can provide advantages to a vessel operator. A device that provides data in an easy to follow format can also provide advantages to the vessel operator. Such advantages may include requiring less time to request and review information, which in turn provides the vessel operator with more time to monitor the marine traffic.
- Described herein are implementations of various technologies for a method and apparatus for providing sonar images. In one implementation, sonar images that are currently being recorded may be displayed on a regular pane. Further, condensed sonar images that had previously been recorded and that are currently being recorded may be displayed on a preview pane, while recording the sonar images.
- Described herein are also implementations of various technologies for an apparatus for displaying sonar images. The apparatus includes one or more processors, a screen, and a memory. The memory has a plurality of executable instructions. When the executable instructions are executed by the one or more processors, the processors may display sonar images that are being recorded at a present time, receive a request to display past sonar images that had been recorded at a previous time, and display the past sonar images while still causing the sonar images that are currently being captured at the present time to be recorded.
- Described herein are also implementations of various technologies for a non-transitory computer readable medium having stored thereon computer-executable instructions which, when executed by a computer, cause the computer to: display sonar images that are currently being recorded on a regular pane; and display condensed sonar images that had previously been recorded and that are currently being recorded on a preview pane, while recording the sonar images.
- In some implementations, a computing system is provided and includes at least one processor, at least one memory, and one or more programs stored in the at least one memory; and means for performing any method disclosed herein.
- In some implementations, an information processing apparatus for use in a computing system is provided, and includes means for performing any method disclosed herein.
- The above referenced summary section is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description section. The summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.
- Implementations of various technologies will hereafter be described with reference to the accompanying drawings. It should be understood, however, that the accompanying drawings illustrate only the various implementations described herein and are not meant to limit the scope of various technologies described herein.
-
FIG. 1 is a schematic diagram of a display device in accordance with various techniques described therein. -
FIG. 2 is a timing diagram in accordance with various techniques described therein. -
FIG. 3 is a schematic diagram of a graphical user interface showing past sonar images in accordance with various techniques described therein. -
FIG. 4 is a flow diagram describing the display of past sonar images in accordance with various techniques described therein. -
FIG. 5 is a schematic diagram of a graphical user interface with a preview window in accordance with various techniques described therein. -
FIG. 6 is a schematic diagram of a graphical user interface showing past sonar images using a preview window in accordance with various techniques described therein. -
FIG. 7 is a flow diagram for showing past sonar images using a preview window in accordance with various techniques described therein. -
FIG. 8 is a flow diagram describing the preview window modes in accordance with various techniques described therein. -
FIG. 9 is a schematic diagram of a computing system in accordance with various techniques described therein. - The discussion below is directed to certain specific implementations. It is to be understood that the discussion below is only for the purpose of enabling a person with ordinary skill in the art to make and use any subject matter defined now or later by the patent “claims” found in any issued patent herein.
- It is specifically intended that the claimed invention not be limited to the implementations and illustrations contained herein, but include modified forms of those implementations including portions of the implementations and combinations of elements of different implementations as come within the scope of the following claims. Nothing in this application is considered critical or essential to the claimed invention unless explicitly indicated as being “critical” or “essential.”
- Reference will now be made in detail to various implementations, examples of which are illustrated in the accompanying drawings and figures. In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the present disclosure. However, it will be apparent to one of ordinary skill in the art that the present disclosure may be practiced without these specific details. In other instances, well-known methods, procedures, components, circuits and networks have not been described in detail so as not to unnecessarily obscure aspects of the implementations.
- It will also be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first object or step could be termed a second object or step, and, similarly, a second object or step could be termed a first object or step, without departing from the scope of the invention. The first object or step, and the second object or step, are both objects or steps, respectively, but they are not to be considered the same object or step.
- The terminology used in the description of the present disclosure herein is for the purpose of describing particular implementations only and is not intended to be limiting of the present disclosure. As used in the description of the present disclosure and the appended claims, the singular forms “a,” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “includes,” “including,” “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components and/or groups thereof.
- As used herein, the term “if” may be construed to mean “when” or “upon” or “in response to determining” or “in response to detecting,” depending on the context. Similarly, the phrase “if it is determined” or “if [a stated condition or event] is detected” may be construed to mean “upon determining” or “in response to determining” or “upon detecting [the stated condition or event]” or “in response to detecting [the stated condition or event],” depending on the context.
- “Alternatively” shall not be construed to only pertain to situations where the number of choices involved is exactly two, but rather refers to another possibility among many other possibilities.
- Additionally, various technologies and techniques described herein include receiving user requests for a number of different operations. In certain instances, the user request for a particular operation will be explicitly described. It shall be understood that “user request” or “user can request” shall also include, but are not limited to, touching the screen, double tapping the screen (tapping the screen twice in rapid succession), pressing a particular physical or virtual button, making a selection from a menu, placement of a cursor at a particular location, stylus pointing, mouse selection, an audible command, as well as the explicit description of the “user request” for the particular operation or how the “user can request”, explicitly, the particular operations.
- In one implementation, various technologies and techniques described herein are directed to a display device having a screen. In one implementation, the screen can display sonar images that are currently being captured as well as sonar images that had previously been captured. The sonar images form a stream of sonar images that may be ordered based on time.
- In another implementation, while the screen displays sonar images that are currently being captured, the display device may provide the user access to older portions of sonar images that were captured considerably long before the currently captured sonar images. For example, the user may access to these older portions by placing their finger on the screen over the sonar data and swiping in the direction of the most recent sonar image. The sonar images on the screen will move according to the movement of the finger. Sonar images on the right (the most recent) will move off the screen, and the older sonar images will appear on the left hand side of the screen. The user can continue this process (swiping the screen and moving the displayed sonar data) until the sonar images from a desired time appear on the screen.
- While the user accesses older sonar images, the display device may continue to record captured sonar images and add them to the sonar image stream. The user can return to viewing the sonar images that are currently being captured. In one implementation, the user can view the sonar images that are currently being captured while the user is viewing older sonar images by swiping the displayed sonar images to the left. When the displayed sonar images are swiped to the left, the oldest sonar images drop off the right hand side and newer sonar images appear on the left hand side. Although various implementations have been described with the most recent sonar images may appear on the right, while older sonar images may appear on the left, in some implementations, the order can be reversed.
- The sonar image stream can be quite lengthy and the user may have to search a long time to view a particular sonar image of interest. In some implementations, the display device may provide a preview of the entire sonar image stream in a smaller region of the screen, e.g., a preview pane. The sonar images in the preview pane may be condensed. However, in certain implementations, the condensed sonar images may be large enough for the user to notice some of the significant characteristics. The user can identify the desired sonar image by spotting the significant characteristics in the condensed sonar image stream. To select the desired sonar image for view in more detail in a regular pane, the user can highlight the desired area in the condensed sonar image stream, and the corresponding sonar image in regular size will appear in the regular pane.
-
FIG. 1 illustrates a block diagram of adisplay device 100 in accordance with various implementations described herein. The display device includes ascreen 105 that provides a graphical user interface for displayingsonar images 110.Sonar images 110 may provide a map of the underlying body of water in the vicinity of a sonar device. Thedisplay device 100 may also include a plurality ofbuttons 112 to facilitate user interaction. The plurality ofbuttons 112 can be physical buttons (shown) or virtual buttons (not shown). Additionally, thedisplay device 100 can include a speaker/microphone 113. - In certain implementations, the
screen 105 can be touch-sensitive or sensitive to the human body heat of a finger. As such, thedisplay device 100 may be able to detect the location of thescreen 105 that is touched by the finger (not shown). In some implementations, thedisplay device 100 may provide acursor 114. Thecursor 114 may be movable by placement of the finger over thecursor 114 and dragging the finger to a desired location. Thedisplay device 100 may cause thecursor 114 to follow the finger. - The
sonar images 110 may be presented as a two-dimensional graphics, where thevertical axis 115 represents depth and thehorizontal axis 120 represents time. The graphical user interface may provide a number of tools to allow the user to obtain more specific information. For example, the user can use thecursor 114 for pointing to a particular location in the graphical user interface. In certain implementations, thedisplay device 100 presents thecursor 114 in response to receiving a user request for thecursor 114. - The graphical user interface can provide specific information about the particular location, such as
depth 130 and estimatedwater temperature 132 at the particular location. The graphical user interface can also provideocean depth 155, estimatedwater temperature 160,latitude 170 andlongitude 175 at a current location of a vessel carrying thedisplay device 100. The graphical user interface can also provide zooming 180 for more detail or more range. - Sonar equipment (not shown) can capture sonar images in real time. The
display device 100 is configured to receive the sonar images and record the sonar images to a recording medium (not shown), such as memory, and also provide the sonar images for display on thescreen 105 in real time. In one implementation, the newly capturedsonar images 110 appear on the rightmost part of the graphic, while theolder sonar images 110 appear towards the left. After some time, theolder sonar images 110 on the left move off of the screen. By convention, the right side has morerecent sonar images 110 while the left side hasolder sonar images 110. In certain implementations, the foregoing can be reversed. - In some implementations, the recording medium can be part of the display device, while in other implementations, the recording medium can be external to the
display device 100. Thedisplay device 100 allows the user to access a previously captured and recorded sonar images. - While the user accesses previously recorded sonar images, the
display device 100 may continue to record the sonar images that the sonar device captures into memory. The sonar images are stored and available for display.FIG. 2 illustrates a timing diagram showing different times that sonar images are recorded and displayed.Axis 200 indicates time, t0 . . . tn, wherein t0 comes first, t1 comes second, t2 comes third, etc. Thesonar images 110 received by the sonar device can be associated with time. For example, at time t4, the user may request thedisplay device 100 to display the sonar images collected at time t0 to t1, which were collected earlier than time t4. The user may review the sonar images collected between time t0 to t1 from t4 to t8 While the user reviews the sonar images at t0 to t1, the sonar images from t4 to t8 are continually being captured and recorded. These sonar images are retrievable for viewing by the user at a later time. For example, at time t8, the user can view sonar images captured at time t6. - The
display device 100 may provide the user with at least two ways of requesting access topast sonar images 110. These include a panning mode and a preview mode. - Reviewing Past Sonar Images
- A. Panning Mode
-
FIG. 3 illustrates a schematic diagram of a graphical user interface showingpast sonar images 110 by panning. When panning, the user places theirfinger 305 onto thescreen 105 over thesonar images 110 and swipes theirfinger 305 in aparticular direction 310. Thesonar images 110 move in the direction of thefinger 305. This is known as panning. Certain portions of thesonar images 110 will no longer appear on the screen, while new portions of the recorded sonar images will appear on thescreen 105. - For example, swiping the
finger 305 to the right provides older sonar images, while swiping thefinger 305 to the left provides more recent sonar images. In one implementation, the data will not move to the left if the mostrecent sonar image 110 is displayed. Additionally, thesonar images 110 will not move to the right once the user reaches theoldest sonar images 110 that are stored. In certain implementations, the user can use one or more of the plurality ofbuttons 112. For example, some of the plurality ofbuttons 112 can be associated with directions. -
FIG. 4 is a flow diagram for reviewingpast sonar images 110 in accordance with various implementations described herein. The flow diagram will be described with references toFIGS. 1-3 . Atblock 405, thedisplay device 100 displayscurrent sonar images 110. - At
block 410, thedisplay device 100 receives a user request to move thesonar images 110 backwards, or to view pastsonar images 110. The user request may include placement of thefinger 305 on thescreen 105 and swiping the finger to the right. For example, at time t4, the user may wish to review thesonar images 110 captured at t0. - At
block 415, when the finger swipe to the right is finished, the desired pastsonar images 110 captured at t0 may be displayed on thescreen 105. In one implementation, thedisplay device 100 displays thepast sonar images 110 in accordance with the user request, while also recording current sonar images. Thedisplay device 100 can display thepast sonar images 110 by receiving finger swipes that cause the display device to continuously pan thesonar images 110. - At
block 420, thedisplay device 100 receives a user request to reviewsonar images 110 that were just recorded while the user was viewingpast sonar images 110 during blocks 410-415. For example, while the user was reviewing past sonar images to-t1 during blocks 410-415 at times t4-t8, thedisplay device 100 was recording sonar images captured from times t4-t8. At time t8, the user may wish to review the sonar images captured at time t6. In one implementation, the user request can include swiping the finger to the left. The user may continue swiping/panning until thedisplay device 100 displays sonar images captured at time t6. - At
block 425, thedisplay device 100 displays thesonar images 110 requested by the user duringblock 420, e.g., sonar images captured at time t6. - At
block 430, while the user is viewing the sonar images captured at time t6, thedisplay device 100 may receive a user request to display the most current sonar images. This user request may be made for example at time t9. The user request can include swiping the screen to the left until the mostrecent sonar images 110, i.e., sonar images captured at time t9, are displayed on thedisplay device 100. Atblock 435, thedisplay device 100 displays the most current sonar images, e.g., t9 . . . tn. - B. Preview Pane
-
FIG. 5 is a schematic of a graphical user interface with apreview pane 505 in accordance with various implementations described herein. In one implementation, thepreview pane 505 displays previously recordedsonar images 110 at a size that is smaller in size to thesonar images 110 displayed in aregular pane 515. In another implementation, thepreview pane 505 can includesonar images 110 covering a longer period of time than theregular pane 515. For example, thepreview pane 505 can include a condensed version of all the sonar images 110 (condensed sonar images) that have been recorded. The recording medium may have a capacity that limits the time period that thesonar images 110 can be stored. For example, the time period can be about ten minutes. In certain implementations, thepreview pane 505 can provide condensessonar images 110 from the most recent hour. Although thepreview pane 505 is described herein as having condensed sonar images, it should be understood that the sonar images may also be in compressed format. - In still another implementation, the
preview pane 505 may include a semi-transparent slidingbox 510 that identifies the portion of thepreview pane 505 representing thesonar images 110 that are shown in theregular pane 515. When theregular pane 515 displays the mostcurrent sonar images 110, the slidingbox 510 is all the way to the right. -
FIG. 6 illustrates viewingpast sonar images 110 using thepreview pane 505 in accordance with various implementations described herein. In one implementation, the condensed sonar images in thepreview pane 505 are smaller than thesonar images 110 in theregular pane 515, but are still large enough for the user to scan and locate notable features. The condensed sonar images in the slidingbox 510 are displayed at regular size in theregular pane 515. In one implementation, moving the slidingbox 510 moves thesonar images 110 in theregular pane 515 to correspond to the condensed sonar images in the slidingbox 510. The user can switch the positions of the slidingbox 510 by placing afinger 305 on the desired condensed sonar images in thepreview pane 505. Upon placing thefinger 305 on the desired condensed sonar images, the slidingbox 510 is displayed around that area and the sonar images in regular size that correspond to the desired condensed sonar images are displayed at theregular pane 515. - In certain implementations, when the user places the
finger 305 on the slidingbox 510 and swipes the slidingbox 510, thesonar images 110 in theregular pane 515 freeze until the user releases the slidingbox 510. When the user releases the slidingbox 510, the sonar images displayed in theregular pane 515 correspond to the condensed sonar images in the slidingbox 510. -
FIG. 7 is a flow diagram for reviewingpast sonar images 110 using thepreview pane 505 in accordance with various implementations described herein. Atblock 705, thedisplay device 100 displayssonar images 110 that are currently captured in aregular pane 515. Atblock 710, thedisplay device 100 displays condensed sonar images that were previously recorded in thepreview pane 505, while displaying currently recordedsonar images 110. Atblock 715, thedisplay device 100 highlights a first portion of the condensed sonar images on thepreview pane 505 with a semi-transparent box, such as slidingbox 510. Atblock 720, thedisplay device 100 displays the sonar images that correspond to the first portion on theregular pane 515 in regular format. - At
block 725, thedisplay device 100 may receive a user request to highlight a second portion of the condensed sonar images on thepreview pane 505. The user request can include receiving finger contact with the slidingbox 510 and swiping the slidingbox 510. Alternatively, the user request can include receiving direct finger contact at the second portion of thepreview pane 505. - At
block 730, thedisplay device 100 highlights the second portion of the condensed sonar images on thepreview pane 505. Atblock 735, thedisplay device 100 displays the sonar images in theregular pane 515 that correspond to the condensed sonar images in the second portion of thepreview pane 505. - Modes of Preview Pane Operation
- Returning to
FIG. 6 , the graphical user interface can include a menu 605 with a “Back” selection 610(1), an “Off” selection 610(2), a “Cursor Only” selection 610(3), and an “Always” selection 610(4). In certain implementations, thedisplay device 100 can be multifunctional and provide additional functions to just sonar image presentation. For example, the “Back” selection 610(1) allows the user to leave the sonar image presentation function and access a higher level menu. - In addition, the
display device 100 can provide thepreview pane 505, based on user selections. For instance, if the user selects the “Off” selection 610(2), thedisplay device 100 does not display thepreview pane 505. The user can also request that thedisplay device 100 always display thepreview pane 505 by selection the “Always” selection 610(4). In some implementations, thepreview pane 505 appears based on whether there is acursor 114. If there is nocursor 114, thepreview pane 505 does not appear. If there is acursor 114 present, thepreview pane 505 appears. The user can request display of the preview pane based on existence of thecursor 114 by selecting the “Cursor Only” selection 610(3). -
FIG. 8 is a flow diagram for determining whether to place thepreview pane 505 or not. Atblock 805, thedisplay device 100 determines whether the mode is “Off”, “Always”, or “Cursor Only”. If atblock 805, thedisplay device 100 determines that the mode is “Off”, atblock 810, thedisplay device 100 does not display thepreview pane 505 and returns to block 805. If atblock 805, thedisplay device 100 determines that the mode is “Always”, atblock 815, thedisplay device 100 displays thepreview pane 505 and returns to block 805. - If at
block 805, thedisplay device 100 determines that thedisplay device 100 is in “Cursor Only” mode, thedisplay device 100 makes a determination whether the cursor is present (block 820). If thecursor 114 is present, thedisplay device 100 displays thepreview pane 505 atblock 815 and returns to block 805. If the cursor is not present atblock 820, thedisplay device 100 does not display the preview pane 505 (block 815). Thedisplay device 100 then returns to block 805. - Computing System
- Implementations of various technologies described herein may be operational with numerous general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the various technologies described herein include, but are not limited to, personal computers, server computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
- The various technologies described herein may be implemented in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that performs particular tasks or implement particular abstract data types. Further, each program module may be implemented in its own way, and all need not be implemented the same way. While program modules may all execute on a single computing system, it should be appreciated that, in some implementations, program modules may be implemented on separate computing systems or devices adapted to communicate with one another. A program module may also be some combination of hardware and software where particular tasks performed by the program module may be done either through hardware, software, or both.
- The various technologies described herein may also be implemented in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network, e.g., by hardwired links, wireless links, or combinations thereof. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
-
FIG. 9 illustrates a schematic diagram of acomputing system 900 in which the various technologies described herein may be incorporated and practiced. Although thecomputing system 900 may be a conventional desktop, a handheld device, personal digital assistant, a server computer, electronic device/instrument, laptop, tablet, or part of a navigation system, marine electronics, or sonar system. It should be noted, however, that other computer system configurations may be used. - The
computing system 900 may include a central processing unit (CPU) 930, asystem memory 926, a graphics processing unit (GPU) 931 and asystem bus 928 that couples various system components including thesystem memory 926 to theCPU 930. Although only one CPU is illustrated inFIG. 9 , it should be understood that in some implementations thecomputing system 900 may include more than one CPU. - The
CPU 930 can include a microprocessor, a microcontroller, a processor, a programmable integrated circuit, or a combination thereof. TheCPU 930 can comprise an off-the-shelf processor such as a Reduced Instruction Set Computer (RISCTM), or a Microprocessor without Interlocked Pipeline Stages (MIPSTM) processor, or a combination thereof. TheCPU 930 can also comprise a proprietary processor. - The
GPU 931 may be a microprocessor specifically designed to manipulate and implement computer graphics. TheCPU 930 may offload work to theGPU 931. TheGPU 931 may have its own graphics memory, and/or may have access to a portion of thesystem memory 926. As with theCPU 930, theGPU 931 may include one or more processing units, and each processing unit may include one or more cores. - The
CPU 930 can provide output data to agraphics processor 931. Thegraphics processor 931 can generate graphical user interfaces that present the output data. Thegraphics processor 931 can also provide objects, such as menus, in the graphical user interface. A user can provide inputs by interacting with the objects. The graphics processor 1531 can receive the inputs from interaction with the objects and provide the inputs to theCPU 930. - The
system bus 928 may be any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. By way of example, and not limitation, such architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus also known as Mezzanine bus. Thesystem memory 926 may include a read only memory (ROM) 912 and a random access memory (RAM) 916. A basic input/output system (BIOS) 914, containing the basic routines that help transfer information between elements within thecomputing system 900, such as during start-up, may be stored in theROM 912. - The
computing system 900 may further include a harddisk drive interface 936 for reading from and writing to ahard disk 950, amagnetic disk drive 952 for reading from and writing to a removablemagnetic disk 956, and anoptical disk drive 954 for reading from and writing to a removableoptical disk 958, such as a CD ROM or other optical media. Thehard disk 950, themagnetic disk drive 952, and theoptical disk drive 954 may be connected to thesystem bus 928 by a harddisk drive interface 936, a magneticdisk drive interface 938, and anoptical drive interface 940, respectively. The drives and their associated computer-readable media may provide nonvolatile storage of computer-readable instructions, data structures, program modules and other data for thecomputing system 900. - Although the
computing system 900 is described herein as having a hard disk, a removablemagnetic disk 956 and a removableoptical disk 958, it should be appreciated by those skilled in the art that thecomputing system 900 may also include other types of computer-readable media that may be accessed by a computer. For example, such computer-readable media may include computer storage media and communication media. Computer storage media may include volatile and non-volatile, and removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules or other data. Computer storage media may further include RAM, ROM, erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other solid state memory technology, CD-ROM, digital versatile disks (DVD), or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by thecomputing system 900. Communication media may embody computer readable instructions, data structures, program modules or other data in a modulated data signal, such as a carrier wave or other transport mechanism and may include any information delivery media. The term “modulated data signal” may mean a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Thecomputing system 900 may also include ahost adapter 933 that connects to astorage device 935 via a small computer system interface (SCSI) bus, a Fiber Channel bus, an eSATA bus, or using any other applicable computer bus interface. Combinations of any of the above may also be included within the scope of computer readable media. - A number of program modules may be stored on the
hard disk 950,magnetic disk 956,optical disk 958,ROM 912 orRAM 916, including anoperating system 918, one ormore application programs 920,control pilot 922, andprogram data 924. In certain implementations, thehard disk 950 may include and a database system. The database system could record, for example, recorded points. Theapplication programs 920 may include various mobile applications (“apps”) and other applications configured to perform various methods and techniques described herein. Theoperating system 918 may be any suitable operating system that may control the operation of a networked personal or server computer. - A user may enter commands and information into the
computing system 900 through input devices such asbuttons 962. Other input devices may include a microphone (not shown). These and other input devices may be connected to theCPU 930 through aserial port interface 942 coupled tosystem bus 928, but may be connected by other interfaces, such as a parallel port, game port or a universal serial bus (USB). - Additionally, the user can enter inputs through a
monitor 934. As noted above, in certain implementations, the screen 905 can be sensitive to touching by afinger 110, while in other implementations, the screen 905 can be sensitive to the body heat from thefinger 110. The screen 905 detects placement and movement of the user's fingers and provides signals that are indicative of such placement and movement to theCPU 930. Additionally, the screen 905 detects user interactions with the objects provided by theGPU 931 and provides signals that are indicative of the user interactions with the objects. Themonitor 934 may also be connected tosystem bus 928 via an interface, such as avideo adapter 932. - Further, the
computing system 900 may operate in a networked environment using logical connections to one or moreremote computers 974. The logical connections may be any connection that is commonplace in offices, enterprise-wide computer networks, intranets, and the Internet, such as local area network (LAN) 976 and a wide area network (WAN) 966. Theremote computers 974 may be another computer, a server computer, a router, a network PC, a peer device or other common network node, and may include many or all of the elements describes above relative to thecomputing system 900. Theremote computers 974 may also each includeapplication programs 970 similar to that of the computer action function. - When using a LAN networking environment, the
computing system 900 may be connected to thelocal network 976 through a network interface oradapter 944. When used in a WAN networking environment, thecomputing system 900 may include awireless router 964 or other means for establishing communication over awide area network 966, such as the Internet. Therouter 964 may be connected to thesystem bus 928 via theserial port interface 942. In a networked environment, program modules depicted relative to thecomputing system 900, or portions thereof, may be stored in a remotememory storage device 972. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between the computers may be used. - In certain implementations, the
computing system 900 can receive sonar signals from a sonar system through thenetwork interface 944. - In certain implementations, the
central processing unit 930,GPU 931,video adapter 932,network interface 944, andsystem memory 926,hard disc 950, monitor 934, and screen 905 can be integrated into a console. - Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
- While the foregoing is directed to implementations of various technologies described herein, other and further implementations may be devised without departing from the basic scope thereof, which may be determined by the claims that follow. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/759,760 US9298079B2 (en) | 2012-07-06 | 2013-02-05 | Sonar preview mode |
PCT/US2013/047645 WO2014008044A1 (en) | 2012-07-06 | 2013-06-25 | Sonar preview mode |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261668965P | 2012-07-06 | 2012-07-06 | |
US13/759,760 US9298079B2 (en) | 2012-07-06 | 2013-02-05 | Sonar preview mode |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140010042A1 true US20140010042A1 (en) | 2014-01-09 |
US9298079B2 US9298079B2 (en) | 2016-03-29 |
Family
ID=49878429
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/759,760 Active 2034-01-16 US9298079B2 (en) | 2012-07-06 | 2013-02-05 | Sonar preview mode |
Country Status (2)
Country | Link |
---|---|
US (1) | US9298079B2 (en) |
WO (1) | WO2014008044A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170115389A1 (en) * | 2009-07-14 | 2017-04-27 | Navico Holding As | Downscan imaging sonar |
US9829573B2 (en) | 2015-07-20 | 2017-11-28 | Navico Holding As | Sonar auto depth range |
US11221403B2 (en) | 2018-05-21 | 2022-01-11 | Navico Holding As | Impact detection devices and methods |
US11249176B2 (en) * | 2018-11-30 | 2022-02-15 | Navico Holding As | Systems and associated methods for monitoring vessel noise level |
US11353566B2 (en) * | 2018-04-26 | 2022-06-07 | Navico Holding As | Sonar transducer having a gyroscope |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10852429B2 (en) | 2014-05-30 | 2020-12-01 | FLIR Belgium BVBA | Sonar history display control systems and methods |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6084565A (en) * | 1995-08-08 | 2000-07-04 | Furuno Electric Company Limited | Image monitoring apparatus |
Family Cites Families (62)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4829493A (en) | 1985-06-14 | 1989-05-09 | Techsonic Industries, Inc. | Sonar fish and bottom finder and display |
US5191341A (en) | 1987-12-01 | 1993-03-02 | Federation Francaise De Voile | System for sea navigation or traffic control/assistance |
US4879697A (en) | 1988-08-05 | 1989-11-07 | Lowrance Electronics, Inc. | Sonar fish finder apparatus providing split-screen display |
US5025423A (en) | 1989-12-21 | 1991-06-18 | At&T Bell Laboratories | Enhanced bottom sonar system |
US6321158B1 (en) | 1994-06-24 | 2001-11-20 | Delorme Publishing Company | Integrated routing/mapping information |
US5537380A (en) | 1995-07-14 | 1996-07-16 | Lowrance Electronics, Inc. | Sonar system having an interactive sonar viewing apparatus and method of configuring same |
US6225984B1 (en) | 1998-05-01 | 2001-05-01 | Hitachi Micro Systems, Inc. | Remote computer interface |
US6388688B1 (en) | 1999-04-06 | 2002-05-14 | Vergics Corporation | Graph-based visual navigation through spatial environments |
US6411283B1 (en) | 1999-05-20 | 2002-06-25 | Micron Technology, Inc. | Computer touch screen adapted to facilitate selection of features at edge of screen |
US7607022B1 (en) | 1999-06-11 | 2009-10-20 | General Instrument Corporation | Configurable encryption/decryption for multiple services support |
US7068774B1 (en) | 2000-02-25 | 2006-06-27 | Harris Corporation | Integrated acd and ivr scripting for call center tracking of calls |
JP4124944B2 (en) | 2000-03-31 | 2008-07-23 | 古野電気株式会社 | Underwater detector |
US6421299B1 (en) | 2000-06-28 | 2002-07-16 | Techsonic Industries, Inc. | Single-transmit, dual-receive sonar |
FR2813684B1 (en) | 2000-09-04 | 2003-01-10 | Jean Baptiste Dumas | DEVICE FOR EXPORTING DATA FROM A DATABASE TO A COMPUTER TERMINAL AND ASSOCIATED METHOD |
US6768942B1 (en) | 2000-09-18 | 2004-07-27 | Navigation Technologies Corp. | Navigation system with decryption functions and secure geographic database |
US7002579B2 (en) | 2001-05-09 | 2006-02-21 | Cadec Corporation | Split screen GPS and electronic tachograph |
US6761692B2 (en) | 2001-06-25 | 2004-07-13 | Eagle Ultrasound As | High frequency and multi frequency band ultrasound transducers based on ceramic films |
US6574554B1 (en) | 2001-12-11 | 2003-06-03 | Garmin Ltd. | System and method for calculating a navigation route based on non-contiguous cartographic map databases |
US20030223733A1 (en) | 2002-05-14 | 2003-12-04 | Intelligent Digital Systems Llc | System and method of processing audio/video data in a remote monitoring system |
US6816782B1 (en) | 2002-10-10 | 2004-11-09 | Garmin Ltd. | Apparatus, systems and methods for navigation data transfer between portable devices |
US6909946B1 (en) | 2002-10-31 | 2005-06-21 | Garmin Ltd. | System and method for wirelessly linking electronic marine components |
US8063540B2 (en) | 2004-03-08 | 2011-11-22 | Emantec As | High frequency ultrasound transducers based on ceramic films |
US7289390B2 (en) | 2004-07-19 | 2007-10-30 | Furuno Electric Company, Limited | Ultrasonic transmitting/receiving apparatus and scanning sonar employing same |
AU2005271581B9 (en) | 2004-08-02 | 2011-01-06 | Johnson Outdoors Inc. | Sonar imaging system for mounting to watercraft |
US7236426B2 (en) | 2004-09-08 | 2007-06-26 | Lowrance Electronics, Inc. | Integrated mapping and audio systems |
US7430461B1 (en) | 2004-10-18 | 2008-09-30 | Navico International Limited | Networking method and network for marine navigation devices |
US20060119585A1 (en) | 2004-12-07 | 2006-06-08 | Skinner David N | Remote control with touchpad and method |
JP4625346B2 (en) | 2005-02-25 | 2011-02-02 | 株式会社リコー | Information processing system |
BRPI0607818A2 (en) | 2005-03-07 | 2009-10-06 | Networks In Motion Inc | method and system for identifying and defining virtual geographic fences |
US7434155B2 (en) | 2005-04-04 | 2008-10-07 | Leitch Technology, Inc. | Icon bar display for video editing system |
US7809175B2 (en) | 2005-07-01 | 2010-10-05 | Hologic, Inc. | Displaying and navigating computer-aided detection results on a review workstation |
US20090099871A1 (en) | 2005-08-09 | 2009-04-16 | Gopal Gadodia | Workflow Oriented Multiscreen Healthcare Information Management System |
US8135504B2 (en) | 2006-04-12 | 2012-03-13 | Craig Summers | Navigational planning and display method for the sailor's dilemma when heading upwind |
US7890867B1 (en) | 2006-06-07 | 2011-02-15 | Adobe Systems Incorporated | Video editing functions displayed on or near video sequences |
US20080126935A1 (en) | 2006-11-08 | 2008-05-29 | Ross James Blomgren | Audio Visual Entertainment System and Method of Operation |
US20080192575A1 (en) | 2007-02-14 | 2008-08-14 | Navico Inc. | Method, Apparatus and Computer Program Product for Providing a Sonar History |
KR101450584B1 (en) | 2007-02-22 | 2014-10-14 | 삼성전자주식회사 | Method for displaying screen in terminal |
US11126321B2 (en) | 2007-09-04 | 2021-09-21 | Apple Inc. | Application menu user interface |
WO2009055918A1 (en) | 2007-11-02 | 2009-05-07 | Marport Canada Inc. | System and method for underwater seismic data acquisition |
US7623667B2 (en) | 2008-01-14 | 2009-11-24 | Apple Inc. | Electronic device accessory with ultrasonic tone generator |
EP2243326B1 (en) | 2008-01-30 | 2018-10-24 | Google LLC | Notification of mobile device events |
US7812667B2 (en) | 2008-03-10 | 2010-10-12 | Qualcomm Incorporated | System and method of enabling a signal processing device in a relatively fast manner to process a low duty cycle signal |
US8305844B2 (en) | 2008-08-07 | 2012-11-06 | Depasqua Louis | Sonar navigation system and method |
US8682576B2 (en) | 2008-12-04 | 2014-03-25 | Verizon Patent And Licensing Inc. | Navigation based on user-defined points and paths |
US7870496B1 (en) | 2009-01-29 | 2011-01-11 | Jahanzeb Ahmed Sherwani | System using touchscreen user interface of a mobile device to remotely control a host computer |
US8601401B2 (en) | 2009-01-30 | 2013-12-03 | Navico Holding As | Method, apparatus and computer program product for synchronizing cursor events |
JP5356874B2 (en) | 2009-03-26 | 2013-12-04 | 古野電気株式会社 | Sailing assistance device |
US20110004839A1 (en) | 2009-07-02 | 2011-01-06 | Derek Cha | User-customized computer display method |
US8300499B2 (en) | 2009-07-14 | 2012-10-30 | Navico, Inc. | Linear and circular downscan imaging sonar |
US8305840B2 (en) | 2009-07-14 | 2012-11-06 | Navico, Inc. | Downscan imaging sonar |
US8643508B2 (en) | 2009-07-26 | 2014-02-04 | Aspen Avionics, Inc. | Avionics device, systems and methods of display |
KR101451999B1 (en) | 2009-07-28 | 2014-10-21 | 삼성전자주식회사 | Data scroll method and apparatus |
WO2011069131A1 (en) | 2009-12-04 | 2011-06-09 | Google Inc. | Presenting real time search results |
US8972903B2 (en) | 2010-07-08 | 2015-03-03 | Apple Inc. | Using gesture to navigate hierarchically ordered user interface screens |
US20120032974A1 (en) | 2010-08-04 | 2012-02-09 | Lynch Phillip C | Method and apparatus for map panning recall |
US8665668B2 (en) | 2010-09-17 | 2014-03-04 | Vivid Engineering, Inc. | Ultrasonic distance measurement controller |
ES2686934T3 (en) | 2011-01-18 | 2018-10-22 | Savant Systems Llc. | Remote control interface that provides operation with raised head and visual response |
JPWO2012144632A1 (en) | 2011-04-20 | 2014-07-28 | Necカシオモバイルコミュニケーションズ株式会社 | Information processing apparatus, information processing method, and program |
KR102023801B1 (en) | 2011-06-05 | 2019-09-20 | 애플 인크. | Systems and methods for displaying notifications received from multiple applications |
US9718530B2 (en) | 2012-04-17 | 2017-08-01 | Garmin Switzerland Gmbh | Marine vessel display system |
US20140013270A1 (en) | 2012-07-06 | 2014-01-09 | Navico Holding As | Multiple Chart Display |
US20140013272A1 (en) | 2012-07-06 | 2014-01-09 | Navico Holding As | Page Editing |
-
2013
- 2013-02-05 US US13/759,760 patent/US9298079B2/en active Active
- 2013-06-25 WO PCT/US2013/047645 patent/WO2014008044A1/en active Application Filing
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6084565A (en) * | 1995-08-08 | 2000-07-04 | Furuno Electric Company Limited | Image monitoring apparatus |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170115389A1 (en) * | 2009-07-14 | 2017-04-27 | Navico Holding As | Downscan imaging sonar |
US9829573B2 (en) | 2015-07-20 | 2017-11-28 | Navico Holding As | Sonar auto depth range |
US11353566B2 (en) * | 2018-04-26 | 2022-06-07 | Navico Holding As | Sonar transducer having a gyroscope |
US11221403B2 (en) | 2018-05-21 | 2022-01-11 | Navico Holding As | Impact detection devices and methods |
US11249176B2 (en) * | 2018-11-30 | 2022-02-15 | Navico Holding As | Systems and associated methods for monitoring vessel noise level |
Also Published As
Publication number | Publication date |
---|---|
US9298079B2 (en) | 2016-03-29 |
WO2014008044A1 (en) | 2014-01-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10789776B2 (en) | Structural modeling using depth sensors | |
KR102027612B1 (en) | Thumbnail-image selection of applications | |
US9298079B2 (en) | Sonar preview mode | |
US10564806B1 (en) | Gesture actions for interface elements | |
US9639186B2 (en) | Multi-touch interface gestures for keyboard and/or mouse inputs | |
US9009594B2 (en) | Content gestures | |
US20180335938A1 (en) | Techniques for image-based search using touch controls | |
EP2287809B1 (en) | Image processing device, image processing method, and computer program | |
US20140013276A1 (en) | Accessing a Marine Electronics Data Menu | |
US10359905B2 (en) | Collaboration with 3D data visualizations | |
US20130222329A1 (en) | Graphical user interface interaction on a touch-sensitive device | |
US10585581B2 (en) | Controlling display object on display screen | |
US20140013270A1 (en) | Multiple Chart Display | |
KR20140104709A (en) | Method for synthesizing images captured by portable terminal, machine-readable storage medium and portable terminal | |
US9442636B2 (en) | Quick split mode | |
US9304612B2 (en) | Off-screen input capture for mobile device | |
US9778824B1 (en) | Bookmark overlays for displayed content | |
CN114578951A (en) | Display device and control method thereof | |
US20150185826A1 (en) | Mapping gestures to virtual functions | |
CN111427505A (en) | Page operating method, device, terminal and storage medium | |
JP5835240B2 (en) | Information processing apparatus, information processing method, and program | |
US9817566B1 (en) | Approaches to managing device functionality | |
WO2017190360A1 (en) | Medical detection system and control method therefor | |
US9122366B2 (en) | Residue indicators | |
CN119002693A (en) | Interactive method, device, electronic equipment, storage medium, system and computer program product executed by electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NAVICO HOLDING AS, NORWAY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:THOMAS, STEVE;COLEMAN, AARON;SIGNING DATES FROM 20131024 TO 20140515;REEL/FRAME:033189/0356 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: GLAS AMERICAS LLC, NEW YORK Free format text: SECURITY INTEREST;ASSIGNOR:NAVICO HOLDING AS;REEL/FRAME:042121/0692 Effective date: 20170331 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: NAVICO HOLDING AS, NORWAY Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:GLAS AMERICAS LLC;REEL/FRAME:057780/0496 Effective date: 20211004 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: NAVICO, INC., OKLAHOMA Free format text: NUNC PRO TUNC ASSIGNMENT;ASSIGNOR:NAVICO HOLDING AS;REEL/FRAME:066338/0107 Effective date: 20230310 |