US20080211915A1 - Scalable system for wide area surveillance - Google Patents

Scalable system for wide area surveillance Download PDF

Info

Publication number
US20080211915A1
US20080211915A1 US12/035,395 US3539508A US2008211915A1 US 20080211915 A1 US20080211915 A1 US 20080211915A1 US 3539508 A US3539508 A US 3539508A US 2008211915 A1 US2008211915 A1 US 2008211915A1
Authority
US
United States
Prior art keywords
logic block
coupled
synchronization
controller
port
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/035,395
Other versions
US8587661B2 (en
Inventor
David L. McCubbrey
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pixel Velocity Inc
Original Assignee
Pixel Velocity Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pixel Velocity Inc filed Critical Pixel Velocity Inc
Priority to US12/035,395 priority Critical patent/US8587661B2/en
Publication of US20080211915A1 publication Critical patent/US20080211915A1/en
Assigned to PIXEL VELOCITY, INC. reassignment PIXEL VELOCITY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MCCUBBREY, DAVID L
Application granted granted Critical
Publication of US8587661B2 publication Critical patent/US8587661B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19608Tracking movement of a target, e.g. by detecting an object predefined as a target, using target direction and or velocity to predict its new position
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19639Details of the system layout
    • G08B13/19641Multiple cameras having overlapping views on a single scene
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources

Definitions

  • This invention relates generally to the surveillance field, and more specifically to a scalable system and method for wide area surveillance.
  • Synchronizing cameras includes obtaining a desired fixed relationship among corresponding significant instants of two or more sensor data streams.
  • the ability to synchronize remote cameras is important when implementing one or more of the following features of a surveillance system: (1) properly collect, fuse, analyze, and interpret sensor data of a target object within a similar view from different sensors, (2) properly collect, align, analyze, and interpret sensor data of a target object across adjacent views of different sensors, (3) quickly adjust or modify the collection of the sensor data based on the analysis and interpretation of adjacent sensor data or other data, and (4) properly interpret sensor data based on a “big picture” perspective of the site.
  • Ethernet which is a standard networking protocol used in local area networks, provides a low-cost method of communicating digital image data from remote cameras.
  • the ability to synchronize the remote cameras over an Ethernet connection is very challenging based on the latencies that depend on the network load, which is variable and unpredictable.
  • the IEEE 1394 standard for a high performance serial bus allows tight synchronization of digital cameras belonging to the same bus.
  • the IEEE 1394 technique has, however, the drawback that cameras belonging to the same bus each must share the total bus bandwidth, so that the technique fails to scale to systems containing more than 2 or 3 cameras.
  • Analog cameras have a concept of “genlock” that allows cameras to be synchronized to a common timing source.
  • the drawback to this approach is that it requires an extra cable to be run between all cameras. This is impractical in systems employing widely distributed cameras, due to installation and cabling costs.
  • Distributed surveillance systems typically aggregate information using a network of PCs. But networked surveillance information introduces unavoidable and uncertain delays due to network routing and latency. This is a significant drawback in situations requiring low latency, such as object tracking. In addition, aggregation of processed results using a network also introduces latency.
  • FIG. 1 is a schematic representation of a controller of the preferred embodiment of the invention.
  • FIG. 2 is a schematic representation of a controller of a variation including a synchronization controller of the preferred embodiment of the invention.
  • FIG. 3 is a schematic representation of coupled controllers of the preferred embodiment of the invention.
  • FIG. 4 is a schematic representation of a system of the preferred embodiment of the invention.
  • FIG. 5 is a schematic representation of a system of the preferred embodiment of the invention including a synchronization controller.
  • FIG. 6 is a schematic representation of a system of the preferred embodiment of the invention including a processor.
  • FIG. 7 is a schematic representation of a network of systems of the preferred embodiment of the invention including a low latency switch.
  • FIG. 8 is a schematic representation of a network of systems of the preferred embodiment of the invention, including a global synchronization controller.
  • FIG. 9 is a schematic representation of a network of systems of the preferred embodiment of the invention, including a split feed bus connection between systems.
  • FIG. 10 is a schematic representation of the lag times in a sensor information aggregation network.
  • FIG. 11 is a schematic representation of an arrangement of sensors with overlapping fields of view.
  • FIG. 12 is a representation of the application of object tracking across multiple cameras.
  • FIG. 13 is a representation of the application of thermal verification with visible imaging.
  • FIG. 14 is a representation of the application of 3-dimensional face matching.
  • FIG. 15 is a flowchart representation of the preferred method of the invention.
  • the invention preferably includes many tightly synchronized acquisition nodes that feed to one or more aggregation nodes through isochronous communication links.
  • both the acquisition nodes and information aggregation nodes preferably contain processing intelligence to optimally distribute the computational loading.
  • This model is highly scalable in both acquisition nodes and layers of aggregation nodes.
  • the combination of low latency, tight synchronization, and distributed computation provided by the model enables many important wide-area surveillance applications, including distributed tracking of moving objects across many views, fusing different image modalities such as visible and InfraRed into products that deliver better information, high dynamic range imaging, adaptive behavior, and 3D imaging of moving objects using multi-look feature correspondence.
  • the invention preferably includes a method and system for collecting information from many remote imagers or sensors, more preferably cameras, that aggregate and process information in real-time to infer events in the environment.
  • a scalable controller is shown in FIGS. 1-3
  • a scalable chassis based subsystem is shown in FIGS. 4-6
  • a scalable low latency multi-chassis system is shown in FIGS. 7-9 .
  • Arrangements and applications of the invention are shown in FIGS. 10-14 .
  • a method of capturing surveillance images is shown in FIG. 15 .
  • a controller 100 for a surveillance system includes ports 103 for coupling to a camera 101 , synchronization logic blocks 105 coupled to the ports 103 , an information aggregation logic block 111 coupled to the camera ports 103 , and an output port 113 coupled to the information aggregation logic block 111 .
  • the preferred embodiment further includes an image sharing logic block 107 coupled to the synchronization logic block 105 of at least two ports 103 , and an image pair-processing logic block 109 coupled to the image sharing logic block 107 .
  • the ports 103 function to connect the controller 100 to a camera 101 .
  • the camera port 103 is preferably a standard IEEE 1394 over CAT-5 port, but may alternatively be a standard Firewire (IEEE 1394b) port, a Serial ATA port, a USB port, an IEEE 1394b over coax connector, or any other suitable port for a connection for a camera 101 to the controller 100 .
  • the output port 113 is preferably a PCI Express port, but may alternatively be an IEEE 1394 port or any other suitable port.
  • the information aggregation logic block 111 preferably functions to aggregate the data analysis from all of the cameras 101 , and transmit the consolidated results to an information bus, or to main memory for a processor 209 for further processing on the synchronized information.
  • the information aggregation logic block 111 is preferably implemented in a Field Programmable gate array, but may alternatively be an ASIC or any other suitable device.
  • the synchronization logic blocks 105 function to synchronize the cameras 101 connected to the controller 100 .
  • the synchronization logic block 105 maintains consistency of the CYCLE_TIME register of the camera 101 to prevent drift over time. This may be accomplished by periodic update of the CYCLE_TIME value, for which the synchronization logic blocks 105 preferably receive a synchronization signal from a synchronization controller 106 .
  • the synchronization logic blocks 105 are preferably implemented in a Field Programmable Gate Array (FPGA), but may alternatively be implemented in an Application Specific Integrated Circuit (ASIC), software on a microprocessor, or any other suitable device.
  • FPGA Field Programmable Gate Array
  • ASIC Application Specific Integrated Circuit
  • the synchronization controller 106 (also known as a timing monarch), functions to generate a synchronization signal.
  • the synchronization controller 106 preferably generates a global time setting signal (also known as a synchronization signal) to ensure consistent values for the IEEE 1394 BUS_TIME and CYCLE_TIME registers of a camera 101 as it is capturing images, such that the image capture and/or subsequent transfer between multiple cameras 101 is synchronized.
  • the synchronization signal preferably establishes a common and consistent value for BUS_TIME and CYCLE_TIME across all connected devices with a timing uncertainty less than half that of the CYCLE_TIME interval.
  • the synchronization controller 106 preferably generates an isochronous frequency signal as the synchronization signal to mark timing cycle starts and periodically re-synchronizes additional controllers 100 or systems 200 that may be connected to the synchronization controller 106 , more preferably the frequency signal is an 8 kHz signal (the CYCLE_TIME in this case is 125 microseconds). Multicast is required for proper coordination of global timing if the cameras 101 are not cross connected (cross connected cameras are shown in FIGS. 1-3 ). In addition to the common re-synchronizer, a global time setting service is provided to guarantee consistent values for each channel's individual BUS_TIME and CYCLE_TIME registers.
  • the controller 100 also includes an image sharing logic block 107 and an image pair-processing logic block 109 . These additional logic blocks preferably provide the controller 100 with dedicated reconfigurable hardware resources for per-image-stream and image-to-image stream processing functions.
  • the image sharing logic block 107 preferably accepts images from at least two synchronized cameras 101 as inputs and preferably outputs the images to an image-pair-processing logic block 109 , but may alternatively use any number of images from any number of cameras 101 , and may connect to any number of image pair-processing logic blocks.
  • An image stream from any channel may be processed on a per-channel and additionally on a channel pair basis, by proper routing of image streams and allocation of processing resources.
  • a stream may be directed to more than one camera-pair processor.
  • the image sharing logic block 107 is preferably implemented in a FPGA, but may alternatively be implemented in an ASIC, a multiplexer chip, or any other suitable implementation.
  • the image pair-processing logic block 109 preferably uses the routed images from the image sharing logic block 107 to perform image processing on pairs of images.
  • the images are captured with partially or fully overlapping fields of view, then the image pair-processing block 109 preferably combines the images, more preferably stitching and/or fusing images together, and/or generating super-resolution images from overlapping areas.
  • the image pair-processing logic block 109 may generate thermal heat signature verification images, preferably by overlaying a thermal image over an optical image.
  • the image pair-processing logic block 109 may track moving objects from one camera field of view to another and simply transmit local track coordinates to the information aggregation logic block 111 .
  • the image pair-processing logic block 109 may combine overlapping images from different perspectives to generate a 3-dimensional image.
  • the image pair-processing logic block 109 is preferably implemented in an FPGA, but may alternatively be an ASIC or any other suitable implementation.
  • a surveillance system 200 includes at least two controllers 100 , a communication channel 215 connected to the controllers 100 , and a chassis 202 supporting the controllers 100 .
  • controllers 100 are preferably identical to the controllers 100 described above (and shown in FIGS. 1-3 ).
  • the controllers 100 are preferably manufactured as a modular device, more preferably as controller cards, and still more preferably as PCI express cards, but may alternatively be any type of controller card, such as IEEE 1394, USB, Serial ATA, or any other suitable device.
  • the controller provides a mechanism for sharing data streams between controllers 100 , implemented as a high speed point-to-point link 115 . Sharing data directly between controllers 100 preserves external PCIe bandwidth, and allows greater flexibility in processing of overlapping fields of view.
  • Multiple controller cards 100 within the same chassis 202 preferably have local links 115 that provide card-to-card channel synchronization and image sharing. As shown in FIG.
  • one controller card 100 may be connected to a synchronization controller 106 that supplies the isochronous timer to all other controller cards 100 in that chassis 202 by means of a series of local card-card links 115 .
  • Each controller card 100 preferably supports connections for at least 4 cameras 101 .
  • the communication channel 215 connected to the controllers 100 preferably functions to enable communication between the controllers 100 .
  • the communication channel 215 is preferably a PCI-Express bus, but may be an IEEE 1394 (Firewire), PCI, ISA, or any other suitable low-latency communication channel.
  • the interface between the controllers 100 and the communication channel 215 is preferably a socketed connection, more preferably a socket that connects to a PCI-express bus, but any suitable connection may be used, including wireless or daisy chained communication channels.
  • the communication channel 215 preferably supports up to 64 controllers 100 .
  • the communication channel 215 may include a synchronization and/or image data channel.
  • the communication channel 215 may connect the controllers 100 to an additional system on a network.
  • the chassis 202 functions to support the controllers 100 , and in the preferred embodiment, to support the communication channel 215 .
  • the chassis 202 also preferably functions to provide power to the controllers 100 .
  • the chassis 202 may be made of aluminum or steel, and preferably supports sixty-four controllers 100 in the form of controller cards, preferably connected to a PCI Express bus.
  • the controllers 100 are preferably fastened to the chassis 202 by a fastener or by any other suitable device or method.
  • a synchronization controller 106 may be connected to the synchronization channel of at least one controller 100 in the system 200 .
  • the synchronization controller 106 preferably uses a protocol built upon an isochronous long-distance interconnect (such as IEEE 1394) by adding a global timing service to keep separate bus trees synchronized within a single chassis 202 .
  • IEEE 1394 an isochronous long-distance interconnect
  • two additional synchronization services are preferably added to ensure consistent local time-keeping for every camera time domain. This is preferably accomplished by designating a single “Timing Monarch”/synchronization channel to provide a continuous source for 8 kHz isochronous cycle starts.
  • the common 8 kHz signal preferably serves to periodically re-synchronize each independent channel connected to a single chassis 202 .
  • a processor 209 is connected to the communication channel 215 , and preferably performs processing functions on the data received from each controller 100 .
  • the controllers 100 preferably forward images or analytic results (or both) to main memory for further processing by a microprocessor 209 .
  • results or images may be forwarded to specialized aggregation processors residing elsewhere on the PCIe bus. This technique provides high bandwidth, low latency cost effective analysis and aggregation of image channels and analysis results within a single chassis 202 .
  • a larger scale system 300 is preferably constructed by connecting multiple (possibly remote) individual systems 200 together.
  • a large scale system 300 for surveillance includes at least two systems 200 and a communication channel 215 connecting the systems 200 .
  • a low latency multi-chassis system 300 using star topology using a low latency switch 319 to connect systems 200 is shown in FIG. 7
  • a serial network topology using a processor 209 and/or a synchronization controller 106 is shown in FIG. 8 .
  • FIG. 9 An aggregation timing diagram is shown in FIG. 10 .
  • the systems 200 are preferably identical to the systems 200 described above (and shown in FIGS. 4-6 ).
  • the communication channel 215 is preferably a PCI Express bus, but may alternatively be IEEE 1394, or any other suitable communication channel.
  • a low latency switch 319 is used to bridge multiple systems 200 together, more preferably bridging PCI express based fan-in tree systems 200 .
  • the low latency switch 319 preferably uses low latency techniques, such as those employed by a PCI express bus switch or Infiniband. These interconnect technologies provide the necessary performance for low latency information aggregation. Multicast is ideally used for proper coordination of global timing if systems 200 are not cross connected (cross connected variants are shown in FIGS. 8-9 ).
  • the basic requirement for multi-chassis synchronization is to be able to establish a common and consistent value for IEEE 1394 BUS_TIME and CYCLE_TIME across all systems 200 , preferably with a timing uncertainty less that half that of the CYCLE_TIME interval (125 microseconds in the preferred embodiment, which uses an 8 kHz synchronization signal). For instance, Infiniband, with a switch latency of only 2.25 microseconds and capability for broadcast transactions can easily meet this requirement. PCI express, without broadcast, must enforce sequential system synchronization activity within a critical time interval.
  • a large scale system 300 includes a processor 209 .
  • the processor 209 preferably functions to process data received from the systems 200 in the network.
  • the network of systems 300 includes a synchronization controller 106 .
  • the processor 209 may also function as a synchronization controller 106 .
  • the communication channel that connects the systems 200 may be at least one split camera feed 317 , providing both time synchronization and image sharing between multiple chassis systems 200 (if geographically separated).
  • This synchronization and/or data channel preferably forms a three-node isochronous bus.
  • the third node may be connected to a remote chassis 202 and may propagate CYCLE_TIME and BUS_TIME to the remote system 200 using appropriate broadcast commands, such as IEEE 1394 CYCLE_START commands.
  • the channel receiving BUS_TIME and CYCLE_TIME information is designated a secondary timing monarch (or a synchronization controller repeater) for the remote system 200 .
  • the synchronization channel preferably synchronizes all other channels in the system contained within the chassis 202 (and any additional systems 200 that may be connected).
  • the general situation for a system encompassing many cameras includes processing nodes (indicated by circles) that perform some amount of analysis before forwarding the results to a higher level in the tree.
  • a conventional implementation of such an architecture would be to allocate a PC to each processing node and link them using an Ethernet TCP/IP network. This approach is capable of scaling and supporting a large number of cameras, but it provides poor efficiency and poor latency. Typical network packet latencies are on the order of 1 millisecond on a local network without intervening routers. Moreover, support for this communication consumes a significant amount of CPU overhead at each of the processing nodes.
  • the network illustrated in FIG. 10 would have a round-trip latency of 8 milliseconds under the conditions of a typical implementation.
  • the method and system of the preferred embodiment preferably uses a processing and communication architecture based on a low cost commodity bus such as PCI Express (PCIe).
  • PCIe achieves node-node latencies under 10 microseconds worst case, making total communication latency lower by a factor of at least 100.
  • PCIe communication protocol is handled primarily in hardware, communication overhead is considerably reduced as well.
  • the cameras are preferably digital video cameras, but may alternatively be still image cameras, thermal cameras, night vision cameras, audio and visual sensors (such as a camera or a microphone), tracking sensors (such as a thermal sensor, a pressure sensor, a RFID detector, or a motion detector), and ambient sensors (such as a rain sensor, a carbon monoxide sensor, a dust sensor, a water or water level sensor, a humidity sensor, a temperature sensor, a light sensor, a methane sensor, a smoke detector, or a wind sensor), or any other suitable cameras or sensors.
  • audio and visual sensors such as a camera or a microphone
  • tracking sensors such as a thermal sensor, a pressure sensor, a RFID detector, or a motion detector
  • ambient sensors such as a rain sensor, a carbon monoxide sensor, a dust sensor, a water or water level sensor, a humidity sensor, a temperature sensor, a light sensor, a methane sensor, a smoke detector, or a wind sensor, or any other suitable cameras or sensors.
  • the sensor arrangement of the preferred embodiment can be varied based on the intended use of the method and system.
  • the cameras are preferably positioned with overlapping fields of view, but may be positioned to capture stereo images, to capture an optical and an infrared image, or any other suitable positioning that may be useful for an imaging application.
  • the method and system may be used to track an object across multiple cameras.
  • the images from the cameras are preferably stitched together to provide panoramic views.
  • the method and system may be used to rapidly identify and track an object. In this variation, an image from a camera is fused with an output from a thermal sensor.
  • the image and the output are preferably focused or pointed in the same general direction.
  • the method and system may be used to generate a 3-D face model.
  • a camera pair is preferably focused or pointed toward the same general area from two different angles.
  • the method and system may be used to generate a 3-D human model.
  • a camera pair is preferably focused or pointed toward the same general area from two different angles.
  • the cameras of the fourth variation are generally pointed downward toward the human objects.
  • the method and system may be used to acquire high dynamic range images by acquiring synchronized and closely aligned views of the same scene, but with different individual camera integration times.
  • a method of the preferred embodiment 1500 includes synchronizing a plurality of cameras S 1510 , capturing images from the synchronized cameras S 1520 , aggregating information from at least two synchronized images S 1530 , and processing the aggregated information S 1540 .
  • the method of the preferred embodiment also includes the steps of distributing synchronized images from at least two cameras to a processor S 1524 , and processing the synchronized images S 1528 .
  • Step S 1510 which recites synchronizing a plurality of cameras, functions to send a synchronization signal to two or more cameras to enable synchronized image capture.
  • the synchronization signal is described above.
  • Step S 1520 which recites capturing images from the synchronized cameras, functions to capture images from the cameras in a time-coordinated fashion, which may be synchronized (such that all images are taken at the same time) or syncopated (such that images are taken at alternating times), or any other time coordinated image capture.
  • Step S 1530 which recites aggregating information from at least two processed synchronized images, functions to consolidate the analysis and/or processed image data from the cameras and/or image processing step and transmit the data to a processor.
  • the aggregation may be performed on a controller scale, a system scale or a wide area network (such as a large corporate network or the internet).
  • Step S 1540 which recites processing the aggregated information, functions to perform processing using more information than step S 1528 , preferably image processing for security or surveillance functions, more preferably object tracking, image stitching, super-resolution, image fusion, heat signature verification, 3-d image processing, facial recognition, or any other suitable image processing application.
  • Step S 1524 which recites distributing synchronized images from at least two cameras to a processor, functions to distribute images captured from the cameras to a processor.
  • Step S 1528 which recites processing the synchronized image pairs, functions to process at least two images captured from the cameras before aggregation in step S 1530 .
  • This step may be performed to distribute a portion of the computational load into a device, such as a controller.
  • the processing of the synchronized images in this step may include stitching, fusing, super-resolution, 3D image processing, object tracking, or any other suitable image processing function that may be performed on the synchronized images.
  • the result of the processing may be new images and/or information about the images.
  • Remote means preferably more than 5 m, and more preferably at least 70 m
  • Multiple cameras means preferably more than 3 cameras, and more preferably 16-256 cameras
  • Synchronized means preferably less than 1 microsecond, and more preferably around 40 nanoseconds
  • Low latency means preferably less than 1 millisecond, and more preferably less than 0.05 milliseconds.

Abstract

According to one embodiment, a controller for a surveillance system includes ports for coupling a camera, synchronization logic blocks coupled to the ports, an information aggregation logic block coupled to the camera ports, and an output port coupled to the information aggregation logic block. According to another embodiment, a method of scaling a surveillance system includes synchronizing a plurality of cameras, capturing images from the synchronized cameras, aggregating at least two processed synchronized images, and processing the aggregated synchronized images.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 60/891,006, filed 21 Feb. 2007, which is incorporated in its entirety by this reference.
  • TECHNICAL FIELD
  • This invention relates generally to the surveillance field, and more specifically to a scalable system and method for wide area surveillance.
  • BACKGROUND
  • Synchronizing cameras includes obtaining a desired fixed relationship among corresponding significant instants of two or more sensor data streams. The ability to synchronize remote cameras is important when implementing one or more of the following features of a surveillance system: (1) properly collect, fuse, analyze, and interpret sensor data of a target object within a similar view from different sensors, (2) properly collect, align, analyze, and interpret sensor data of a target object across adjacent views of different sensors, (3) quickly adjust or modify the collection of the sensor data based on the analysis and interpretation of adjacent sensor data or other data, and (4) properly interpret sensor data based on a “big picture” perspective of the site.
  • Ethernet, which is a standard networking protocol used in local area networks, provides a low-cost method of communicating digital image data from remote cameras. The ability to synchronize the remote cameras over an Ethernet connection, however, is very challenging based on the latencies that depend on the network load, which is variable and unpredictable. Unlike Ethernet, the IEEE 1394 standard for a high performance serial bus allows tight synchronization of digital cameras belonging to the same bus. The IEEE 1394 technique has, however, the drawback that cameras belonging to the same bus each must share the total bus bandwidth, so that the technique fails to scale to systems containing more than 2 or 3 cameras.
  • Analog cameras have a concept of “genlock” that allows cameras to be synchronized to a common timing source. The drawback to this approach is that it requires an extra cable to be run between all cameras. This is impractical in systems employing widely distributed cameras, due to installation and cabling costs. Distributed surveillance systems typically aggregate information using a network of PCs. But networked surveillance information introduces unavoidable and uncertain delays due to network routing and latency. This is a significant drawback in situations requiring low latency, such as object tracking. In addition, aggregation of processed results using a network also introduces latency.
  • Thus, there is a need in the surveillance field to create a scalable method and system for wide area surveillance that achieves tight synchronization between a large number of remote cameras, and achieves very scalable high speed and low latency processing at multiple information aggregation levels. This invention provides such new and useful scalable method and system.
  • BRIEF DESCRIPTION OF THE FIGURES
  • FIG. 1 is a schematic representation of a controller of the preferred embodiment of the invention.
  • FIG. 2 is a schematic representation of a controller of a variation including a synchronization controller of the preferred embodiment of the invention.
  • FIG. 3 is a schematic representation of coupled controllers of the preferred embodiment of the invention.
  • FIG. 4 is a schematic representation of a system of the preferred embodiment of the invention.
  • FIG. 5 is a schematic representation of a system of the preferred embodiment of the invention including a synchronization controller.
  • FIG. 6 is a schematic representation of a system of the preferred embodiment of the invention including a processor.
  • FIG. 7 is a schematic representation of a network of systems of the preferred embodiment of the invention including a low latency switch.
  • FIG. 8 is a schematic representation of a network of systems of the preferred embodiment of the invention, including a global synchronization controller.
  • FIG. 9 is a schematic representation of a network of systems of the preferred embodiment of the invention, including a split feed bus connection between systems.
  • FIG. 10 is a schematic representation of the lag times in a sensor information aggregation network.
  • FIG. 11 is a schematic representation of an arrangement of sensors with overlapping fields of view.
  • FIG. 12 is a representation of the application of object tracking across multiple cameras.
  • FIG. 13 is a representation of the application of thermal verification with visible imaging.
  • FIG. 14 is a representation of the application of 3-dimensional face matching.
  • FIG. 15 is a flowchart representation of the preferred method of the invention.
  • DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The following description of the preferred embodiments of the invention is not intended to limit the invention to these preferred embodiments, but rather to enable any person skilled in the art to make and use this invention.
  • The invention preferably includes many tightly synchronized acquisition nodes that feed to one or more aggregation nodes through isochronous communication links. In the preferred embodiment both the acquisition nodes and information aggregation nodes preferably contain processing intelligence to optimally distribute the computational loading. This model is highly scalable in both acquisition nodes and layers of aggregation nodes. In particular, the combination of low latency, tight synchronization, and distributed computation provided by the model enables many important wide-area surveillance applications, including distributed tracking of moving objects across many views, fusing different image modalities such as visible and InfraRed into products that deliver better information, high dynamic range imaging, adaptive behavior, and 3D imaging of moving objects using multi-look feature correspondence.
  • In the preferred embodiment, the invention preferably includes a method and system for collecting information from many remote imagers or sensors, more preferably cameras, that aggregate and process information in real-time to infer events in the environment. A scalable controller is shown in FIGS. 1-3, a scalable chassis based subsystem is shown in FIGS. 4-6, and a scalable low latency multi-chassis system is shown in FIGS. 7-9. Arrangements and applications of the invention are shown in FIGS. 10-14. A method of capturing surveillance images is shown in FIG. 15.
  • As shown in FIGS. 1-3, a controller 100 for a surveillance system includes ports 103 for coupling to a camera 101, synchronization logic blocks 105 coupled to the ports 103, an information aggregation logic block 111 coupled to the camera ports 103, and an output port 113 coupled to the information aggregation logic block 111. The preferred embodiment further includes an image sharing logic block 107 coupled to the synchronization logic block 105 of at least two ports 103, and an image pair-processing logic block 109 coupled to the image sharing logic block 107.
  • The ports 103 function to connect the controller 100 to a camera 101. The camera port 103 is preferably a standard IEEE 1394 over CAT-5 port, but may alternatively be a standard Firewire (IEEE 1394b) port, a Serial ATA port, a USB port, an IEEE 1394b over coax connector, or any other suitable port for a connection for a camera 101 to the controller 100. The output port 113 is preferably a PCI Express port, but may alternatively be an IEEE 1394 port or any other suitable port.
  • The information aggregation logic block 111 preferably functions to aggregate the data analysis from all of the cameras 101, and transmit the consolidated results to an information bus, or to main memory for a processor 209 for further processing on the synchronized information. The information aggregation logic block 111 is preferably implemented in a Field Programmable gate array, but may alternatively be an ASIC or any other suitable device.
  • The synchronization logic blocks 105 function to synchronize the cameras 101 connected to the controller 100. The synchronization logic block 105 maintains consistency of the CYCLE_TIME register of the camera 101 to prevent drift over time. This may be accomplished by periodic update of the CYCLE_TIME value, for which the synchronization logic blocks 105 preferably receive a synchronization signal from a synchronization controller 106. The synchronization logic blocks 105 are preferably implemented in a Field Programmable Gate Array (FPGA), but may alternatively be implemented in an Application Specific Integrated Circuit (ASIC), software on a microprocessor, or any other suitable device.
  • The synchronization controller 106 (also known as a timing monarch), functions to generate a synchronization signal. The synchronization controller 106 preferably generates a global time setting signal (also known as a synchronization signal) to ensure consistent values for the IEEE 1394 BUS_TIME and CYCLE_TIME registers of a camera 101 as it is capturing images, such that the image capture and/or subsequent transfer between multiple cameras 101 is synchronized. The synchronization signal preferably establishes a common and consistent value for BUS_TIME and CYCLE_TIME across all connected devices with a timing uncertainty less than half that of the CYCLE_TIME interval. The synchronization controller 106 preferably generates an isochronous frequency signal as the synchronization signal to mark timing cycle starts and periodically re-synchronizes additional controllers 100 or systems 200 that may be connected to the synchronization controller 106, more preferably the frequency signal is an 8 kHz signal (the CYCLE_TIME in this case is 125 microseconds). Multicast is required for proper coordination of global timing if the cameras 101 are not cross connected (cross connected cameras are shown in FIGS. 1-3). In addition to the common re-synchronizer, a global time setting service is provided to guarantee consistent values for each channel's individual BUS_TIME and CYCLE_TIME registers.
  • In the preferred embodiment of the invention, as shown in FIGS. 1-3, the controller 100 also includes an image sharing logic block 107 and an image pair-processing logic block 109. These additional logic blocks preferably provide the controller 100 with dedicated reconfigurable hardware resources for per-image-stream and image-to-image stream processing functions.
  • The image sharing logic block 107 preferably accepts images from at least two synchronized cameras 101 as inputs and preferably outputs the images to an image-pair-processing logic block 109, but may alternatively use any number of images from any number of cameras 101, and may connect to any number of image pair-processing logic blocks. An image stream from any channel may be processed on a per-channel and additionally on a channel pair basis, by proper routing of image streams and allocation of processing resources. A stream may be directed to more than one camera-pair processor. The image sharing logic block 107 is preferably implemented in a FPGA, but may alternatively be implemented in an ASIC, a multiplexer chip, or any other suitable implementation.
  • The image pair-processing logic block 109 preferably uses the routed images from the image sharing logic block 107 to perform image processing on pairs of images. Preferably, the images are captured with partially or fully overlapping fields of view, then the image pair-processing block 109 preferably combines the images, more preferably stitching and/or fusing images together, and/or generating super-resolution images from overlapping areas. In one variation, if a thermal image is captured from a camera 101, the image pair-processing logic block 109 may generate thermal heat signature verification images, preferably by overlaying a thermal image over an optical image. In another variation, the image pair-processing logic block 109 may track moving objects from one camera field of view to another and simply transmit local track coordinates to the information aggregation logic block 111. In yet another variation, the image pair-processing logic block 109 may combine overlapping images from different perspectives to generate a 3-dimensional image. The image pair-processing logic block 109 is preferably implemented in an FPGA, but may alternatively be an ASIC or any other suitable implementation.
  • As shown in FIGS. 4-6, a surveillance system 200 includes at least two controllers 100, a communication channel 215 connected to the controllers 100, and a chassis 202 supporting the controllers 100.
  • Except as noted below, the controllers 100 are preferably identical to the controllers 100 described above (and shown in FIGS. 1-3). The controllers 100 are preferably manufactured as a modular device, more preferably as controller cards, and still more preferably as PCI express cards, but may alternatively be any type of controller card, such as IEEE 1394, USB, Serial ATA, or any other suitable device. In one variation the controller provides a mechanism for sharing data streams between controllers 100, implemented as a high speed point-to-point link 115. Sharing data directly between controllers 100 preserves external PCIe bandwidth, and allows greater flexibility in processing of overlapping fields of view. Multiple controller cards 100 within the same chassis 202 preferably have local links 115 that provide card-to-card channel synchronization and image sharing. As shown in FIG. 5, one controller card 100 may be connected to a synchronization controller 106 that supplies the isochronous timer to all other controller cards 100 in that chassis 202 by means of a series of local card-card links 115. Each controller card 100 preferably supports connections for at least 4 cameras 101.
  • The communication channel 215 connected to the controllers 100 preferably functions to enable communication between the controllers 100. The communication channel 215 is preferably a PCI-Express bus, but may be an IEEE 1394 (Firewire), PCI, ISA, or any other suitable low-latency communication channel. The interface between the controllers 100 and the communication channel 215 is preferably a socketed connection, more preferably a socket that connects to a PCI-express bus, but any suitable connection may be used, including wireless or daisy chained communication channels. The communication channel 215 preferably supports up to 64 controllers 100. In one variation, the communication channel 215 may include a synchronization and/or image data channel. In another variation, the communication channel 215 may connect the controllers 100 to an additional system on a network.
  • The chassis 202 functions to support the controllers 100, and in the preferred embodiment, to support the communication channel 215. The chassis 202 also preferably functions to provide power to the controllers 100. The chassis 202 may be made of aluminum or steel, and preferably supports sixty-four controllers 100 in the form of controller cards, preferably connected to a PCI Express bus. The controllers 100 are preferably fastened to the chassis 202 by a fastener or by any other suitable device or method.
  • In one variation of the preferred embodiment, as shown in FIG. 5, a synchronization controller 106 (as previously described) may be connected to the synchronization channel of at least one controller 100 in the system 200. The synchronization controller 106 preferably uses a protocol built upon an isochronous long-distance interconnect (such as IEEE 1394) by adding a global timing service to keep separate bus trees synchronized within a single chassis 202. Unlike conventional IEEE 1394, two additional synchronization services are preferably added to ensure consistent local time-keeping for every camera time domain. This is preferably accomplished by designating a single “Timing Monarch”/synchronization channel to provide a continuous source for 8 kHz isochronous cycle starts. The common 8 kHz signal preferably serves to periodically re-synchronize each independent channel connected to a single chassis 202.
  • In another variation of the preferred embodiment, as shown in FIG. 6, a processor 209 is connected to the communication channel 215, and preferably performs processing functions on the data received from each controller 100. In this variation, the controllers 100 preferably forward images or analytic results (or both) to main memory for further processing by a microprocessor 209. Alternatively, results or images may be forwarded to specialized aggregation processors residing elsewhere on the PCIe bus. This technique provides high bandwidth, low latency cost effective analysis and aggregation of image channels and analysis results within a single chassis 202.
  • As shown in FIGS. 7-10, a larger scale system 300 is preferably constructed by connecting multiple (possibly remote) individual systems 200 together. As shown in FIGS. 7-10, a large scale system 300 for surveillance includes at least two systems 200 and a communication channel 215 connecting the systems 200. A low latency multi-chassis system 300 using star topology using a low latency switch 319 to connect systems 200 is shown in FIG. 7, and a serial network topology using a processor 209 and/or a synchronization controller 106 is shown in FIG. 8. A variation of connecting systems 200 using a shared camera bus 317 is shown in FIG. 9. An aggregation timing diagram is shown in FIG. 10.
  • The systems 200 are preferably identical to the systems 200 described above (and shown in FIGS. 4-6). The communication channel 215 is preferably a PCI Express bus, but may alternatively be IEEE 1394, or any other suitable communication channel.
  • In one variation a low latency switch 319 is used to bridge multiple systems 200 together, more preferably bridging PCI express based fan-in tree systems 200. The low latency switch 319 preferably uses low latency techniques, such as those employed by a PCI express bus switch or Infiniband. These interconnect technologies provide the necessary performance for low latency information aggregation. Multicast is ideally used for proper coordination of global timing if systems 200 are not cross connected (cross connected variants are shown in FIGS. 8-9). The basic requirement for multi-chassis synchronization is to be able to establish a common and consistent value for IEEE 1394 BUS_TIME and CYCLE_TIME across all systems 200, preferably with a timing uncertainty less that half that of the CYCLE_TIME interval (125 microseconds in the preferred embodiment, which uses an 8 kHz synchronization signal). For instance, Infiniband, with a switch latency of only 2.25 microseconds and capability for broadcast transactions can easily meet this requirement. PCI express, without broadcast, must enforce sequential system synchronization activity within a critical time interval.
  • In one variation, as shown in FIG. 8, a large scale system 300 includes a processor 209. The processor 209 preferably functions to process data received from the systems 200 in the network. In another variation, also shown in FIG. 8, the network of systems 300 includes a synchronization controller 106. The processor 209 may also function as a synchronization controller 106.
  • In one variation, as shown in FIG. 9, the communication channel that connects the systems 200 may be at least one split camera feed 317, providing both time synchronization and image sharing between multiple chassis systems 200 (if geographically separated). This synchronization and/or data channel preferably forms a three-node isochronous bus. The third node may be connected to a remote chassis 202 and may propagate CYCLE_TIME and BUS_TIME to the remote system 200 using appropriate broadcast commands, such as IEEE 1394 CYCLE_START commands. In the remote system 200, the channel receiving BUS_TIME and CYCLE_TIME information is designated a secondary timing monarch (or a synchronization controller repeater) for the remote system 200. In this variant, the synchronization channel preferably synchronizes all other channels in the system contained within the chassis 202 (and any additional systems 200 that may be connected).
  • As shown in FIG. 10, the general situation for a system encompassing many cameras includes processing nodes (indicated by circles) that perform some amount of analysis before forwarding the results to a higher level in the tree. A conventional implementation of such an architecture would be to allocate a PC to each processing node and link them using an Ethernet TCP/IP network. This approach is capable of scaling and supporting a large number of cameras, but it provides poor efficiency and poor latency. Typical network packet latencies are on the order of 1 millisecond on a local network without intervening routers. Moreover, support for this communication consumes a significant amount of CPU overhead at each of the processing nodes. The network illustrated in FIG. 10 would have a round-trip latency of 8 milliseconds under the conditions of a typical implementation. In contrast, the method and system of the preferred embodiment preferably uses a processing and communication architecture based on a low cost commodity bus such as PCI Express (PCIe). PCIe achieves node-node latencies under 10 microseconds worst case, making total communication latency lower by a factor of at least 100. In addition, since the PCIe communication protocol is handled primarily in hardware, communication overhead is considerably reduced as well.
  • The cameras are preferably digital video cameras, but may alternatively be still image cameras, thermal cameras, night vision cameras, audio and visual sensors (such as a camera or a microphone), tracking sensors (such as a thermal sensor, a pressure sensor, a RFID detector, or a motion detector), and ambient sensors (such as a rain sensor, a carbon monoxide sensor, a dust sensor, a water or water level sensor, a humidity sensor, a temperature sensor, a light sensor, a methane sensor, a smoke detector, or a wind sensor), or any other suitable cameras or sensors.
  • As shown in FIGS. 11-14, the sensor arrangement of the preferred embodiment can be varied based on the intended use of the method and system. As shown in FIG. 11, the cameras are preferably positioned with overlapping fields of view, but may be positioned to capture stereo images, to capture an optical and an infrared image, or any other suitable positioning that may be useful for an imaging application. In a first variation, as shown in FIG. 12, the method and system may be used to track an object across multiple cameras. In this variation, the images from the cameras are preferably stitched together to provide panoramic views. In a second variation, as shown in FIG. 13, the method and system may be used to rapidly identify and track an object. In this variation, an image from a camera is fused with an output from a thermal sensor. In this variation, the image and the output are preferably focused or pointed in the same general direction. In a third variation, as shown in FIG. 14, the method and system may be used to generate a 3-D face model. In this variation, a camera pair is preferably focused or pointed toward the same general area from two different angles. In a fourth variation (not shown), the method and system may be used to generate a 3-D human model. In this variation, like the third variation, a camera pair is preferably focused or pointed toward the same general area from two different angles. Unlike the third variation, however, the cameras of the fourth variation are generally pointed downward toward the human objects. In a fifth variation (not shown) the method and system may be used to acquire high dynamic range images by acquiring synchronized and closely aligned views of the same scene, but with different individual camera integration times.
  • As shown in FIG. 15, a method of the preferred embodiment 1500 includes synchronizing a plurality of cameras S1510, capturing images from the synchronized cameras S1520, aggregating information from at least two synchronized images S1530, and processing the aggregated information S1540. As shown in FIG. 15, the method of the preferred embodiment also includes the steps of distributing synchronized images from at least two cameras to a processor S1524, and processing the synchronized images S1528.
  • Step S1510, which recites synchronizing a plurality of cameras, functions to send a synchronization signal to two or more cameras to enable synchronized image capture. The synchronization signal is described above.
  • Step S1520, which recites capturing images from the synchronized cameras, functions to capture images from the cameras in a time-coordinated fashion, which may be synchronized (such that all images are taken at the same time) or syncopated (such that images are taken at alternating times), or any other time coordinated image capture.
  • Step S1530, which recites aggregating information from at least two processed synchronized images, functions to consolidate the analysis and/or processed image data from the cameras and/or image processing step and transmit the data to a processor. The aggregation may be performed on a controller scale, a system scale or a wide area network (such as a large corporate network or the internet).
  • Step S1540, which recites processing the aggregated information, functions to perform processing using more information than step S1528, preferably image processing for security or surveillance functions, more preferably object tracking, image stitching, super-resolution, image fusion, heat signature verification, 3-d image processing, facial recognition, or any other suitable image processing application.
  • Step S1524, which recites distributing synchronized images from at least two cameras to a processor, functions to distribute images captured from the cameras to a processor.
  • Step S1528, which recites processing the synchronized image pairs, functions to process at least two images captured from the cameras before aggregation in step S1530. This step may be performed to distribute a portion of the computational load into a device, such as a controller. The processing of the synchronized images in this step may include stitching, fusing, super-resolution, 3D image processing, object tracking, or any other suitable image processing function that may be performed on the synchronized images. The result of the processing may be new images and/or information about the images.
  • Within this document, the following terms are hereby defined: “Remote” means preferably more than 5 m, and more preferably at least 70 m; “Multiple cameras” means preferably more than 3 cameras, and more preferably 16-256 cameras; “Synchronized” means preferably less than 1 microsecond, and more preferably around 40 nanoseconds; and “Low latency” means preferably less than 1 millisecond, and more preferably less than 0.05 milliseconds.
  • As a person skilled in the art will recognize from the previous detailed description and from the figures and claims, modifications and changes can be made to the preferred embodiments of the invention without departing from the scope of this invention defined in the following claims.

Claims (25)

1. A controller for a surveillance system, comprising:
a first port for coupling a camera;
a second port for coupling a camera;
a first synchronization logic block coupled to the first port;
a second synchronization logic block coupled to the second port;
an information aggregation logic block coupled to the camera ports; and
an output port coupled to the information aggregation logic block.
2. The system of claim 1, further comprising an image sharing logic block coupled to the synchronization logic block of the first port and the synchronization logic block of the second port; and an image pair-processing logic block coupled to the image sharing logic block and coupled to the information aggregation logic block.
3. The controller of claim 1, further comprising a synchronization controller coupled to the synchronization logic blocks.
4. A system for surveillance, comprising;
a first controller according to the controller of claim 1;
a second controller according to the controller of claim 1;
a communication channel coupled to the controllers; and
a chassis supporting the controllers.
5. The system of claim 4, wherein an output of the image sharing logic block coupled to the first port on the first controller is coupled to an image sharing logic block connecting to the first port on the second controller.
6. The system of claim 4, wherein a synchronization logic block on the first controller is coupled to a synchronization logic block on the second controller.
7. The system of claim 4, wherein the communication channel is one selected from the group consisting of PCI-Express and IEEE 1394.
8. The system of claim 6, further comprising a synchronization controller coupled to the synchronization service logic block of the first controller.
9. The system of claim 8, further comprising a synchronization channel in the communication channel, wherein the synchronization channel on the communication channel is coupled to at least two synchronization logic blocks on a controller.
10. The system of claim 4, further comprising a processor connected to the communication channel.
11. A network of systems for surveillance, comprising;
a first system according to the system of claim 4;
a second system according to the system of claim 4; and
a communication channel coupling the first system and the second system.
12. The system of claim 11, further comprising a low latency switch coupled to the first system and the second system.
13. The system of claim 11, further comprising a synchronization controller.
14. The system of claim 11, further comprising a processor coupled to the communication channel.
15. The system of claim 11, further comprising a plurality of cameras connected to the camera ports on the controllers.
16. The system of claim 15, wherein two of the plurality of cameras have at least partially overlapping fields of view.
17. A method comprising:
synchronizing a plurality of cameras;
capturing images from the synchronized cameras;
aggregating information from at least two synchronized images; and
processing the aggregated information.
18. The method of claim 17, further comprising the step of distributing synchronized images from at least two cameras to a processor.
19. The method of claim 18, further comprising the step of processing the synchronized image pairs.
20. The method of claim 17, wherein the step of synchronizing a plurality of cameras includes providing a continuous source for isochronous cycle start times.
21. The method of claim 20, wherein the step of synchronizing a plurality of cameras includes setting a global time to provide consistent values for the bus_time and cycle_time registers on a communication channel.
22. The method of claim 18, wherein the step of processing the aggregated processed synchronized images includes one application selected from the group consisting of combining optical images with thermal images to produce thermal signature verified images, combining images to create a 3-dimensional image, and tracking objects across multiple camera fields of view.
23. A network of systems for surveillance, comprising:
at least two subsystems, wherein each subsystem includes:
at least two controllers for a surveillance system wherein each controller includes a first port for coupling a camera, a second port for coupling a camera, a first synchronization logic block coupled to the first port, a second synchronization logic block coupled to the second port, an image sharing logic block coupled to the synchronization logic block of the first port and the synchronization logic block of the second port, an image pair-processing logic block coupled to the image sharing logic block, an information aggregation logic block coupled to the output of the image pair-processing logic block, and an output port coupled to the information aggregation logic block;
a communication channel coupled to the output port of the controllers; and
a chassis supporting the controllers;
a low latency switch coupled to the first system and the second system;
a processor coupled to the communication channel; and
a synchronization controller coupled to a synchronization channel on the communication channel coupled to at least two synchronization logic blocks on a controller, wherein the synchronization controller outputs a continuous source for isochronous cycle start times onto the synchronization channel.
24. The system of claim 23, further comprising a plurality of cameras connected to the camera ports on the controllers.
25. The system of claim 24, wherein two of the plurality of cameras have at least partially overlapping fields of view.
US12/035,395 2007-02-21 2008-02-21 Scalable system for wide area surveillance Expired - Fee Related US8587661B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/035,395 US8587661B2 (en) 2007-02-21 2008-02-21 Scalable system for wide area surveillance

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US89100607P 2007-02-21 2007-02-21
US12/035,395 US8587661B2 (en) 2007-02-21 2008-02-21 Scalable system for wide area surveillance

Publications (2)

Publication Number Publication Date
US20080211915A1 true US20080211915A1 (en) 2008-09-04
US8587661B2 US8587661B2 (en) 2013-11-19

Family

ID=39710750

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/035,395 Expired - Fee Related US8587661B2 (en) 2007-02-21 2008-02-21 Scalable system for wide area surveillance

Country Status (5)

Country Link
US (1) US8587661B2 (en)
JP (1) JP2010519860A (en)
GB (1) GB2459602B (en)
IL (1) IL200553A0 (en)
WO (1) WO2008103850A2 (en)

Cited By (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080036864A1 (en) * 2006-08-09 2008-02-14 Mccubbrey David System and method for capturing and transmitting image data streams
US20080148227A1 (en) * 2002-05-17 2008-06-19 Mccubbrey David L Method of partitioning an algorithm between hardware and software
US20080151049A1 (en) * 2006-12-14 2008-06-26 Mccubbrey David L Gaming surveillance system and method of extracting metadata from multiple synchronized cameras
US20080292140A1 (en) * 2007-05-22 2008-11-27 Stephen Jeffrey Morris Tracking people and objects using multiple live and recorded surveillance camera video feeds
US20090086023A1 (en) * 2007-07-18 2009-04-02 Mccubbrey David L Sensor system including a configuration of the sensor as a virtual sensor device
US20100046810A1 (en) * 2008-08-20 2010-02-25 Fujitsu Limited Fingerprint image acquiring device, fingerprint authenticating apparatus, fingerprint image acquiring method, and fingerprint authenticating method
US20100128928A1 (en) * 2008-11-27 2010-05-27 Sony Corporation Image processing apparatus, image processing method, and program
US20100128446A1 (en) * 2008-11-21 2010-05-27 Bosch Security Systems, Inc. Security system including modular ring housing
US20110115909A1 (en) * 2009-11-13 2011-05-19 Sternberg Stanley R Method for tracking an object through an environment across multiple cameras
US20120083314A1 (en) * 2010-09-30 2012-04-05 Ng Hock M Multimedia Telecommunication Apparatus With Motion Tracking
US20120134282A1 (en) * 2010-11-30 2012-05-31 Nokia Corporation Method and apparatus for selecting devices to form a community
US8278779B2 (en) 2011-02-07 2012-10-02 General Electric Company System and method for providing redundant power to a device
US20120265917A1 (en) * 2009-11-09 2012-10-18 Imec Data transferring device
US8300981B1 (en) * 2007-02-28 2012-10-30 Fluke Corporation System and method for analyzing multiple thermal images
US8339457B1 (en) 2007-02-28 2012-12-25 Fluke Corporation Systems and methods for time-shared memory access in a thermal imaging device
US20130135468A1 (en) * 2010-08-16 2013-05-30 Korea Research Institute Of Standards And Science Camera tracing and surveillance system and method for security using thermal image coordinate
US8515127B2 (en) 2010-07-28 2013-08-20 International Business Machines Corporation Multispectral detection of personal attributes for video surveillance
US8532390B2 (en) 2010-07-28 2013-09-10 International Business Machines Corporation Semantic parsing of objects in video
US20130250134A1 (en) * 2012-02-19 2013-09-26 Jack J. McCauley System and Methods for Wireless Remote Control over Cameras
US8587661B2 (en) * 2007-02-21 2013-11-19 Pixel Velocity, Inc. Scalable system for wide area surveillance
US20130329947A1 (en) * 2012-06-06 2013-12-12 Etron Technology, Inc. Image capturing method for image recognition and system thereof
US8754925B2 (en) 2010-09-30 2014-06-17 Alcatel Lucent Audio source locator and tracker, a method of directing a camera to view an audio source and a video conferencing terminal
US20140218612A1 (en) * 2013-02-06 2014-08-07 Robert Bosch Gmbh Multiple field-of-view video streaming
US8917909B2 (en) 2012-06-04 2014-12-23 International Business Machines Corporation Surveillance including a modified video data stream
US9008487B2 (en) 2011-12-06 2015-04-14 Alcatel Lucent Spatial bookmarking
US9071626B2 (en) 2008-10-03 2015-06-30 Vidsys, Inc. Method and apparatus for surveillance system peering
US9087386B2 (en) 2012-11-30 2015-07-21 Vidsys, Inc. Tracking people and objects using multiple live and recorded surveillance camera video feeds
US9134399B2 (en) 2010-07-28 2015-09-15 International Business Machines Corporation Attribute-based person tracking across multiple cameras
US9183560B2 (en) 2010-05-28 2015-11-10 Daniel H. Abelow Reality alternate
US9294716B2 (en) 2010-04-30 2016-03-22 Alcatel Lucent Method and system for controlling an imaging system
US20170264882A1 (en) * 2016-03-10 2017-09-14 Dell Products, Lp System and method of generating a temperature map
US9955209B2 (en) 2010-04-14 2018-04-24 Alcatel-Lucent Usa Inc. Immersive viewer, a method of providing scenes on a display and an immersive viewing system
US20180176425A1 (en) * 2016-12-21 2018-06-21 Shanghai Xiaoyi Technology Co., Ltd. Method and system for configuring cameras to capture images
US20180220069A1 (en) * 2015-07-21 2018-08-02 Hangzhou Hikvision Digital Technology Co., Ltd. Camera and integrated circuit board
EP3275183A4 (en) * 2015-03-27 2018-10-03 Intel Corporation Technologies for controlling user access to image sensors of a camera device
US10094662B1 (en) 2017-03-28 2018-10-09 Trimble Inc. Three-dimension position and heading solution
US20190043003A1 (en) * 2017-08-07 2019-02-07 Standard Cognition, Corp Predicting inventory events using foreground/background processing
US20190052800A1 (en) * 2016-02-29 2019-02-14 Nokia Technologies Oy Adaptive control of image capture parameters in virtual reality cameras
US20190156277A1 (en) * 2017-08-07 2019-05-23 Standard Cognition, Corp Realtime inventory location management using deep learning
US10300573B2 (en) 2017-05-24 2019-05-28 Trimble Inc. Measurement, layout, marking, firestop stick
US10332089B1 (en) * 2015-03-31 2019-06-25 Amazon Technologies, Inc. Data synchronization system
US10341618B2 (en) 2017-05-24 2019-07-02 Trimble Inc. Infrastructure positioning camera system
US10339670B2 (en) 2017-08-29 2019-07-02 Trimble Inc. 3D tool tracking and positioning using cameras
US10347008B2 (en) 2017-08-14 2019-07-09 Trimble Inc. Self positioning camera system to 3D CAD/BIM model
US10406645B2 (en) 2017-05-24 2019-09-10 Trimble Inc. Calibration approach for camera placement
US10424342B2 (en) 2010-07-28 2019-09-24 International Business Machines Corporation Facilitating people search in video surveillance
US10445694B2 (en) 2017-08-07 2019-10-15 Standard Cognition, Corp. Realtime inventory tracking using deep learning
US10474991B2 (en) 2017-08-07 2019-11-12 Standard Cognition, Corp. Deep learning-based store realograms
US10650545B2 (en) 2017-08-07 2020-05-12 Standard Cognition, Corp. Systems and methods to check-in shoppers in a cashier-less store
US10853965B2 (en) 2017-08-07 2020-12-01 Standard Cognition, Corp Directional impression analysis using deep learning
US10997747B2 (en) 2019-05-09 2021-05-04 Trimble Inc. Target positioning with bundle adjustment
US11002541B2 (en) 2019-07-23 2021-05-11 Trimble Inc. Target positioning with electronic distance measuring and bundle adjustment
US11200692B2 (en) 2017-08-07 2021-12-14 Standard Cognition, Corp Systems and methods to check-in shoppers in a cashier-less store
US11232575B2 (en) * 2019-04-18 2022-01-25 Standard Cognition, Corp Systems and methods for deep learning-based subject persistence
US11232687B2 (en) 2017-08-07 2022-01-25 Standard Cognition, Corp Deep learning-based shopper statuses in a cashier-less store
US11250376B2 (en) 2017-08-07 2022-02-15 Standard Cognition, Corp Product correlation analysis using deep learning
US11303853B2 (en) 2020-06-26 2022-04-12 Standard Cognition, Corp. Systems and methods for automated design of camera placement and cameras arrangements for autonomous checkout
US11361468B2 (en) 2020-06-26 2022-06-14 Standard Cognition, Corp. Systems and methods for automated recalibration of sensors for autonomous checkout
US11402271B2 (en) * 2017-06-03 2022-08-02 Teledyne Flir, Llc Extensible architecture for surveillance and targeting imaging systems and methods
US20220417055A1 (en) * 2019-09-19 2022-12-29 Telefonaktiebolaget Lm Ericsson (Publ) Low latency small message communication channel using remote write
US11551079B2 (en) 2017-03-01 2023-01-10 Standard Cognition, Corp. Generating labeled training images for use in training a computational neural network for object or action recognition
US11610142B2 (en) * 2019-05-28 2023-03-21 Ati Technologies Ulc Safety monitor for image misclassification
US11790682B2 (en) 2017-03-10 2023-10-17 Standard Cognition, Corp. Image analysis using neural networks for pose and action identification

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9497380B1 (en) 2013-02-15 2016-11-15 Red.Com, Inc. Dense field imaging
RU2523858C1 (en) * 2013-10-15 2014-07-27 Вячеслав Михайлович Смелков Device for panoramic television surveillance
WO2015098110A1 (en) * 2013-12-27 2015-07-02 パナソニックIpマネジメント株式会社 Imaging device, imaging system, and imaging method
US20160093181A1 (en) 2014-09-26 2016-03-31 Motorola Solutions, Inc Method and apparatus for generating a super-resolved image from multiple unsynchronized cameras
TWI554976B (en) 2014-11-17 2016-10-21 財團法人工業技術研究院 Surveillance systems and image processing methods thereof
US10643076B2 (en) 2016-07-01 2020-05-05 International Business Machines Corporation Counterfeit detection
TWI599989B (en) 2016-11-29 2017-09-21 財團法人工業技術研究院 Image processing method and image system for transportation
US10297059B2 (en) 2016-12-21 2019-05-21 Motorola Solutions, Inc. Method and image processor for sending a combined image to human versus machine consumers
CN108712626B (en) * 2018-03-22 2020-08-04 苏州艾微视图像科技有限公司 Multichannel image acquisition system
US10411793B1 (en) 2018-06-20 2019-09-10 Goodrich Corporation Data transport and time synchronization for ISR systems

Citations (106)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5307168A (en) * 1991-03-29 1994-04-26 Sony Electronics, Inc. Method and apparatus for synchronizing two cameras
US5452239A (en) * 1993-01-29 1995-09-19 Quickturn Design Systems, Inc. Method of removing gated clocks from the clock nets of a netlist for timing sensitive implementation of the netlist in a hardware emulation system
US5623304A (en) * 1989-09-28 1997-04-22 Matsushita Electric Industrial Co., Ltd. CCTV system using multiplexed signals to reduce required cables
US5841439A (en) * 1994-07-22 1998-11-24 Monash University Updating graphical objects based on object validity periods
US5912980A (en) * 1995-07-13 1999-06-15 Hunke; H. Martin Target acquisition and tracking
US6006276A (en) * 1996-10-31 1999-12-21 Sensormatic Electronics Corporation Enhanced video data compression in intelligent video information management system
US6064398A (en) * 1993-09-10 2000-05-16 Geovector Corporation Electro-optic vision systems
US6086629A (en) * 1997-12-04 2000-07-11 Xilinx, Inc. Method for design implementation of routing in an FPGA using placement directives such as local outputs and virtual buffers
US6097429A (en) * 1997-08-01 2000-08-01 Esco Electronics Corporation Site control unit for video security system
US6202164B1 (en) * 1998-07-02 2001-03-13 Advanced Micro Devices, Inc. Data rate synchronization by frame rate adjustment
US6301695B1 (en) * 1999-01-14 2001-10-09 Xilinx, Inc. Methods to securely configure an FPGA using macro markers
US20010046316A1 (en) * 2000-02-21 2001-11-29 Naoki Miyano Image synthesis apparatus
US6370677B1 (en) * 1996-05-07 2002-04-09 Xilinx, Inc. Method and system for maintaining hierarchy throughout the integrated circuit design process
US6373851B1 (en) * 1998-07-23 2002-04-16 F.R. Aleman & Associates, Inc. Ethernet based network to control electronic devices
US6396535B1 (en) * 1999-02-16 2002-05-28 Mitsubishi Electric Research Laboratories, Inc. Situation awareness system
US6438737B1 (en) * 2000-02-15 2002-08-20 Intel Corporation Reconfigurable logic for a computer
US6457164B1 (en) * 1998-03-27 2002-09-24 Xilinx, Inc. Hetergeneous method for determining module placement in FPGAs
US20030025599A1 (en) * 2001-05-11 2003-02-06 Monroe David A. Method and apparatus for collecting, sending, archiving and retrieving motion video and still images and notification of detected events
US6526563B1 (en) * 2000-07-13 2003-02-25 Xilinx, Inc. Method for improving area in reduced programmable logic devices
US20030052966A1 (en) * 2000-09-06 2003-03-20 Marian Trinkel Synchronization of a stereoscopic camera
US20030062997A1 (en) * 1999-07-20 2003-04-03 Naidoo Surendra N. Distributed monitoring for a video security system
US6557156B1 (en) * 1997-08-28 2003-04-29 Xilinx, Inc. Method of configuring FPGAS for dynamically reconfigurable computing
US20030086300A1 (en) * 2001-04-06 2003-05-08 Gareth Noyes FPGA coprocessing system
US6561600B1 (en) * 2000-09-13 2003-05-13 Rockwell Collins In-flight entertainment LCD monitor housing multi-purpose latch
US20030095711A1 (en) * 2001-11-16 2003-05-22 Stmicroelectronics, Inc. Scalable architecture for corresponding multiple video streams at frame rate
US20030098913A1 (en) * 2001-11-29 2003-05-29 Lighting Innovation & Services Co., Ltd. Digital swift video controller system
US20030101426A1 (en) * 2001-11-27 2003-05-29 Terago Communications, Inc. System and method for providing isolated fabric interface in high-speed network switching and routing platforms
US20030160980A1 (en) * 2001-09-12 2003-08-28 Martin Olsson Graphics engine for high precision lithography
US20030174203A1 (en) * 2000-07-19 2003-09-18 Junichi Takeno Image converter for providing flicker-free stereoscopic image based on top-down split frame sequential suitable for computer communications
US6625743B1 (en) * 1998-07-02 2003-09-23 Advanced Micro Devices, Inc. Method for synchronizing generation and consumption of isochronous data
US20030193577A1 (en) * 2002-03-07 2003-10-16 Jorg Doring Multiple video camera surveillance system
US20030217364A1 (en) * 2002-05-17 2003-11-20 Polanek Edward L. System handling video, control signals and power
US6668312B2 (en) * 2001-12-21 2003-12-23 Celoxica Ltd. System, method, and article of manufacture for dynamically profiling memory transfers in a program
US20040061780A1 (en) * 2002-09-13 2004-04-01 Huffman David A. Solid-state video surveillance system
US20040061774A1 (en) * 2002-04-10 2004-04-01 Wachtel Robert A. Digital imaging system using overlapping images to formulate a seamless composite image and implemented using either a digital imaging sensor array
US20040095374A1 (en) * 2002-11-14 2004-05-20 Nebojsa Jojic System and method for automatically learning flexible sprites in video layers
US6754882B1 (en) * 2002-02-22 2004-06-22 Xilinx, Inc. Method and system for creating a customized support package for an FPGA-based system-on-chip (SoC)
US6757304B1 (en) * 1999-01-27 2004-06-29 Sony Corporation Method and apparatus for data communication and storage wherein a IEEE1394/firewire clock is synchronized to an ATM network clock
US6760063B1 (en) * 1996-04-08 2004-07-06 Canon Kabushiki Kaisha Camera control apparatus and method
US20040130620A1 (en) * 2002-11-12 2004-07-08 Buehler Christopher J. Method and system for tracking and behavioral monitoring of multiple objects moving through multiple fields-of-view
US20040135885A1 (en) * 2002-10-16 2004-07-15 George Hage Non-intrusive sensor and method
US6785352B1 (en) * 1999-02-19 2004-08-31 Nokia Mobile Phones Ltd. Method and circuit arrangement for implementing inter-system synchronization in a multimode device
US6798344B2 (en) * 2002-07-08 2004-09-28 James Otis Faulkner Security alarm system and method with realtime streaming video
US20040233983A1 (en) * 2003-05-20 2004-11-25 Marconi Communications, Inc. Security system
US20040240542A1 (en) * 2002-02-06 2004-12-02 Arie Yeredor Method and apparatus for video frame sequence-based object tracking
US20040252193A1 (en) * 2003-06-12 2004-12-16 Higgins Bruce E. Automated traffic violation monitoring and reporting system with combined video and still-image data
US20040252194A1 (en) * 2003-06-16 2004-12-16 Yung-Ting Lin Linking zones for object tracking and camera handoff
US20040263621A1 (en) * 2001-09-14 2004-12-30 Guo Chun Biao Customer service counter/checkpoint registration system with video/image capturing, indexing, retrieving and black list matching function
US20050025313A1 (en) * 2003-06-19 2005-02-03 Wachtel Robert A. Digital imaging system for creating a wide-angle image from multiple narrow angle images
US20050047646A1 (en) * 2003-08-27 2005-03-03 Nebojsa Jojic System and method for fast on-line learning of transformed hidden Markov models
US20050073685A1 (en) * 2003-10-03 2005-04-07 Olympus Corporation Image processing apparatus and method for processing images
US20050073585A1 (en) * 2003-09-19 2005-04-07 Alphatech, Inc. Tracking systems and methods
US6894809B2 (en) * 2002-03-01 2005-05-17 Orasee Corp. Multiple angle display produced from remote optical sensing devices
US20050165995A1 (en) * 2001-03-15 2005-07-28 Italtel S.P.A. System of distributed microprocessor interfaces toward macro-cell based designs implemented as ASIC or FPGA bread boarding and relative COMMON BUS protocol
US20050185053A1 (en) * 2004-02-23 2005-08-25 Berkey Thomas F. Motion targeting system and method
US20050190263A1 (en) * 2000-11-29 2005-09-01 Monroe David A. Multiple video display configurations and remote control of multiple video signals transmitted to a monitoring station over a network
US20050212918A1 (en) * 2004-03-25 2005-09-29 Bill Serra Monitoring system and method
US6970183B1 (en) * 2000-06-14 2005-11-29 E-Watch, Inc. Multimedia surveillance and monitoring system including network configuration
US20050275721A1 (en) * 2004-06-14 2005-12-15 Yusuke Ishii Monitor system for monitoring suspicious object
US6985620B2 (en) * 2000-03-07 2006-01-10 Sarnoff Corporation Method of pose estimation and model refinement for video representation of a three dimensional scene
US20060007242A1 (en) * 2004-07-08 2006-01-12 Microsoft Corporation Matching digital information flow to a human perception system
US20060020990A1 (en) * 2004-07-22 2006-01-26 Mceneaney Ian P System and method for selectively providing video of travel destinations
US20060028552A1 (en) * 2004-07-28 2006-02-09 Manoj Aggarwal Method and apparatus for stereo, multi-camera tracking and RF and video track fusion
US7015954B1 (en) * 1999-08-09 2006-03-21 Fuji Xerox Co., Ltd. Automatic video system using multiple cameras
US20060117356A1 (en) * 2004-12-01 2006-06-01 Microsoft Corporation Interactive montages of sprites for indexing and summarizing video
US7073158B2 (en) * 2002-05-17 2006-07-04 Pixel Velocity, Inc. Automated system for designing and developing field programmable gate arrays
US20060149829A1 (en) * 2004-12-31 2006-07-06 Ta-Chiun Kuan Monitor system
US20060171453A1 (en) * 2005-01-04 2006-08-03 Rohlfing Thomas R Video surveillance system
US20060174302A1 (en) * 2005-02-01 2006-08-03 Bryan Mattern Automated remote monitoring system for construction sites
US20060187305A1 (en) * 2002-07-01 2006-08-24 Trivedi Mohan M Digital processing of video images
US20060192839A1 (en) * 2005-02-25 2006-08-31 Samsung Electronics Co., Ltd. Method and apparatus for measuring optical density of image printed on medium
US20060197839A1 (en) * 2005-03-07 2006-09-07 Senior Andrew W Automatic multiscale image acquisition from a steerable camera
US7106374B1 (en) * 1999-04-05 2006-09-12 Amherst Systems, Inc. Dynamically reconfigurable vision system
US20060215765A1 (en) * 2005-03-25 2006-09-28 Cherng-Daw Hwang Split screen video in a multimedia communication system
US20060227138A1 (en) * 2005-04-05 2006-10-12 Nissan Motor Co., Ltd. Image processing device and method
US20060252554A1 (en) * 2005-05-03 2006-11-09 Tangam Technologies Inc. Gaming object position analysis and tracking
US20060252521A1 (en) * 2005-05-03 2006-11-09 Tangam Technologies Inc. Table game tracking
US20070024706A1 (en) * 2005-08-01 2007-02-01 Brannon Robert H Jr Systems and methods for providing high-resolution regions-of-interest
US20070058717A1 (en) * 2005-09-09 2007-03-15 Objectvideo, Inc. Enhanced processing for scanning video
US20070065002A1 (en) * 2005-02-18 2007-03-22 Laurence Marzell Adaptive 3D image modelling system and apparatus and method therefor
US20070098001A1 (en) * 2005-10-04 2007-05-03 Mammen Thomas PCI express to PCI express based low latency interconnect scheme for clustering systems
US20070104328A1 (en) * 2005-11-04 2007-05-10 Sunplus Technology Co., Ltd. Image signal processing device
US7231065B2 (en) * 2004-03-15 2007-06-12 Embarcadero Systems Corporation Method and apparatus for controlling cameras and performing optical character recognition of container code and chassis code
US20070183770A1 (en) * 2004-12-21 2007-08-09 Katsuji Aoki Camera terminal and imaging zone adjusting apparatus
US20070250898A1 (en) * 2006-03-28 2007-10-25 Object Video, Inc. Automatic extraction of secondary video streams
US20070247525A1 (en) * 2004-06-01 2007-10-25 L-3 Comminications Corporation Video Flashlight/Vision Alert
US20070258009A1 (en) * 2004-09-30 2007-11-08 Pioneer Corporation Image Processing Device, Image Processing Method, and Image Processing Program
US20070279494A1 (en) * 2004-04-16 2007-12-06 Aman James A Automatic Event Videoing, Tracking And Content Generation
US20080019566A1 (en) * 2006-07-21 2008-01-24 Wolfgang Niem Image-processing device, surveillance system, method for establishing a scene reference image, and computer program
US20080036864A1 (en) * 2006-08-09 2008-02-14 Mccubbrey David System and method for capturing and transmitting image data streams
US20080060034A1 (en) * 2006-02-13 2008-03-06 Geoffrey Egnal System and method to combine multiple video streams
US20080074494A1 (en) * 2006-09-26 2008-03-27 Harris Corporation Video Surveillance System Providing Tracking of a Moving Object in a Geospatial Model and Related Methods
US20080096372A1 (en) * 2006-10-23 2008-04-24 Interuniversitair Microelektronica Centrum (Imec) Patterning of doped poly-silicon gates
US20080133767A1 (en) * 2006-11-22 2008-06-05 Metis Enterprise Technologies Llc Real-time multicast peer-to-peer video streaming platform
US7386833B2 (en) * 2002-09-04 2008-06-10 Mentor Graphics Corp. Polymorphic computational system and method in signals intelligence analysis
US20080151049A1 (en) * 2006-12-14 2008-06-26 Mccubbrey David L Gaming surveillance system and method of extracting metadata from multiple synchronized cameras
US7394916B2 (en) * 2003-02-10 2008-07-01 Activeye, Inc. Linking tracked objects that undergo temporary occlusion
US7451410B2 (en) * 2002-05-17 2008-11-11 Pixel Velocity Inc. Stackable motherboard and related sensor systems
US20080297587A1 (en) * 2007-05-31 2008-12-04 Kurtz Andrew F Multi-camera residential communication system
US20090051682A1 (en) * 2003-08-15 2009-02-26 Werner Gerhard Lonsing Method and apparatus for producing composite images which contain virtual objects
US7511764B2 (en) * 2002-07-24 2009-03-31 Alan Neal Cooper Digital camera synchronization
US20090086023A1 (en) * 2007-07-18 2009-04-02 Mccubbrey David L Sensor system including a configuration of the sensor as a virtual sensor device
US7551203B2 (en) * 2002-11-29 2009-06-23 Fujitsu Limited Picture inputting apparatus using high-resolution image pickup device to acquire low-resolution whole pictures and high-resolution partial pictures
US7620266B2 (en) * 2005-01-20 2009-11-17 International Business Machines Corporation Robust and efficient foreground analysis for real-time video surveillance
US20110115909A1 (en) * 2009-11-13 2011-05-19 Sternberg Stanley R Method for tracking an object through an environment across multiple cameras
US8063929B2 (en) * 2007-05-31 2011-11-22 Eastman Kodak Company Managing scene transitions for video communication

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003041411A1 (en) 2001-11-08 2003-05-15 Revolution Company, Llc Video system and methods for operating a video system
GB2459602B (en) * 2007-02-21 2011-09-21 Pixel Velocity Inc Scalable system for wide area surveillance

Patent Citations (115)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5623304A (en) * 1989-09-28 1997-04-22 Matsushita Electric Industrial Co., Ltd. CCTV system using multiplexed signals to reduce required cables
US5307168A (en) * 1991-03-29 1994-04-26 Sony Electronics, Inc. Method and apparatus for synchronizing two cameras
US5452239A (en) * 1993-01-29 1995-09-19 Quickturn Design Systems, Inc. Method of removing gated clocks from the clock nets of a netlist for timing sensitive implementation of the netlist in a hardware emulation system
US6064398A (en) * 1993-09-10 2000-05-16 Geovector Corporation Electro-optic vision systems
US5841439A (en) * 1994-07-22 1998-11-24 Monash University Updating graphical objects based on object validity periods
US5912980A (en) * 1995-07-13 1999-06-15 Hunke; H. Martin Target acquisition and tracking
US6760063B1 (en) * 1996-04-08 2004-07-06 Canon Kabushiki Kaisha Camera control apparatus and method
US6370677B1 (en) * 1996-05-07 2002-04-09 Xilinx, Inc. Method and system for maintaining hierarchy throughout the integrated circuit design process
US6006276A (en) * 1996-10-31 1999-12-21 Sensormatic Electronics Corporation Enhanced video data compression in intelligent video information management system
US6097429A (en) * 1997-08-01 2000-08-01 Esco Electronics Corporation Site control unit for video security system
US6557156B1 (en) * 1997-08-28 2003-04-29 Xilinx, Inc. Method of configuring FPGAS for dynamically reconfigurable computing
US6086629A (en) * 1997-12-04 2000-07-11 Xilinx, Inc. Method for design implementation of routing in an FPGA using placement directives such as local outputs and virtual buffers
US6457164B1 (en) * 1998-03-27 2002-09-24 Xilinx, Inc. Hetergeneous method for determining module placement in FPGAs
US6202164B1 (en) * 1998-07-02 2001-03-13 Advanced Micro Devices, Inc. Data rate synchronization by frame rate adjustment
US6625743B1 (en) * 1998-07-02 2003-09-23 Advanced Micro Devices, Inc. Method for synchronizing generation and consumption of isochronous data
US6373851B1 (en) * 1998-07-23 2002-04-16 F.R. Aleman & Associates, Inc. Ethernet based network to control electronic devices
US6301695B1 (en) * 1999-01-14 2001-10-09 Xilinx, Inc. Methods to securely configure an FPGA using macro markers
US6757304B1 (en) * 1999-01-27 2004-06-29 Sony Corporation Method and apparatus for data communication and storage wherein a IEEE1394/firewire clock is synchronized to an ATM network clock
US6396535B1 (en) * 1999-02-16 2002-05-28 Mitsubishi Electric Research Laboratories, Inc. Situation awareness system
US6785352B1 (en) * 1999-02-19 2004-08-31 Nokia Mobile Phones Ltd. Method and circuit arrangement for implementing inter-system synchronization in a multimode device
US7106374B1 (en) * 1999-04-05 2006-09-12 Amherst Systems, Inc. Dynamically reconfigurable vision system
US20030062997A1 (en) * 1999-07-20 2003-04-03 Naidoo Surendra N. Distributed monitoring for a video security system
US7015954B1 (en) * 1999-08-09 2006-03-21 Fuji Xerox Co., Ltd. Automatic video system using multiple cameras
US6438737B1 (en) * 2000-02-15 2002-08-20 Intel Corporation Reconfigurable logic for a computer
US20010046316A1 (en) * 2000-02-21 2001-11-29 Naoki Miyano Image synthesis apparatus
US7072504B2 (en) * 2000-02-21 2006-07-04 Sharp Kabushiki Kaisha Image synthesis apparatus
US6985620B2 (en) * 2000-03-07 2006-01-10 Sarnoff Corporation Method of pose estimation and model refinement for video representation of a three dimensional scene
US6970183B1 (en) * 2000-06-14 2005-11-29 E-Watch, Inc. Multimedia surveillance and monitoring system including network configuration
US6526563B1 (en) * 2000-07-13 2003-02-25 Xilinx, Inc. Method for improving area in reduced programmable logic devices
US20030174203A1 (en) * 2000-07-19 2003-09-18 Junichi Takeno Image converter for providing flicker-free stereoscopic image based on top-down split frame sequential suitable for computer communications
US20030052966A1 (en) * 2000-09-06 2003-03-20 Marian Trinkel Synchronization of a stereoscopic camera
US6561600B1 (en) * 2000-09-13 2003-05-13 Rockwell Collins In-flight entertainment LCD monitor housing multi-purpose latch
US20050190263A1 (en) * 2000-11-29 2005-09-01 Monroe David A. Multiple video display configurations and remote control of multiple video signals transmitted to a monitoring station over a network
US20050165995A1 (en) * 2001-03-15 2005-07-28 Italtel S.P.A. System of distributed microprocessor interfaces toward macro-cell based designs implemented as ASIC or FPGA bread boarding and relative COMMON BUS protocol
US20030086300A1 (en) * 2001-04-06 2003-05-08 Gareth Noyes FPGA coprocessing system
US20030025599A1 (en) * 2001-05-11 2003-02-06 Monroe David A. Method and apparatus for collecting, sending, archiving and retrieving motion video and still images and notification of detected events
US20030160980A1 (en) * 2001-09-12 2003-08-28 Martin Olsson Graphics engine for high precision lithography
US20040263621A1 (en) * 2001-09-14 2004-12-30 Guo Chun Biao Customer service counter/checkpoint registration system with video/image capturing, indexing, retrieving and black list matching function
US20030095711A1 (en) * 2001-11-16 2003-05-22 Stmicroelectronics, Inc. Scalable architecture for corresponding multiple video streams at frame rate
US7054491B2 (en) * 2001-11-16 2006-05-30 Stmicroelectronics, Inc. Scalable architecture for corresponding multiple video streams at frame rate
US20030101426A1 (en) * 2001-11-27 2003-05-29 Terago Communications, Inc. System and method for providing isolated fabric interface in high-speed network switching and routing platforms
US20030098913A1 (en) * 2001-11-29 2003-05-29 Lighting Innovation & Services Co., Ltd. Digital swift video controller system
US6668312B2 (en) * 2001-12-21 2003-12-23 Celoxica Ltd. System, method, and article of manufacture for dynamically profiling memory transfers in a program
US20040240542A1 (en) * 2002-02-06 2004-12-02 Arie Yeredor Method and apparatus for video frame sequence-based object tracking
US6754882B1 (en) * 2002-02-22 2004-06-22 Xilinx, Inc. Method and system for creating a customized support package for an FPGA-based system-on-chip (SoC)
US6894809B2 (en) * 2002-03-01 2005-05-17 Orasee Corp. Multiple angle display produced from remote optical sensing devices
US20030193577A1 (en) * 2002-03-07 2003-10-16 Jorg Doring Multiple video camera surveillance system
US7215364B2 (en) * 2002-04-10 2007-05-08 Panx Imaging, Inc. Digital imaging system using overlapping images to formulate a seamless composite image and implemented using either a digital imaging sensor array
US20040061774A1 (en) * 2002-04-10 2004-04-01 Wachtel Robert A. Digital imaging system using overlapping images to formulate a seamless composite image and implemented using either a digital imaging sensor array
US20030217364A1 (en) * 2002-05-17 2003-11-20 Polanek Edward L. System handling video, control signals and power
US20080148227A1 (en) * 2002-05-17 2008-06-19 Mccubbrey David L Method of partitioning an algorithm between hardware and software
US7073158B2 (en) * 2002-05-17 2006-07-04 Pixel Velocity, Inc. Automated system for designing and developing field programmable gate arrays
US7587699B2 (en) * 2002-05-17 2009-09-08 Pixel Velocity, Inc. Automated system for designing and developing field programmable gate arrays
US7451410B2 (en) * 2002-05-17 2008-11-11 Pixel Velocity Inc. Stackable motherboard and related sensor systems
US20060187305A1 (en) * 2002-07-01 2006-08-24 Trivedi Mohan M Digital processing of video images
US6798344B2 (en) * 2002-07-08 2004-09-28 James Otis Faulkner Security alarm system and method with realtime streaming video
US7511764B2 (en) * 2002-07-24 2009-03-31 Alan Neal Cooper Digital camera synchronization
US7386833B2 (en) * 2002-09-04 2008-06-10 Mentor Graphics Corp. Polymorphic computational system and method in signals intelligence analysis
US20040061780A1 (en) * 2002-09-13 2004-04-01 Huffman David A. Solid-state video surveillance system
US20040135885A1 (en) * 2002-10-16 2004-07-15 George Hage Non-intrusive sensor and method
US20040130620A1 (en) * 2002-11-12 2004-07-08 Buehler Christopher J. Method and system for tracking and behavioral monitoring of multiple objects moving through multiple fields-of-view
US20070024635A1 (en) * 2002-11-14 2007-02-01 Microsoft Corporation Modeling variable illumination in an image sequence
US20040095374A1 (en) * 2002-11-14 2004-05-20 Nebojsa Jojic System and method for automatically learning flexible sprites in video layers
US20070104383A1 (en) * 2002-11-14 2007-05-10 Microsoft Corporation Stabilization of objects within a video sequence
US7551203B2 (en) * 2002-11-29 2009-06-23 Fujitsu Limited Picture inputting apparatus using high-resolution image pickup device to acquire low-resolution whole pictures and high-resolution partial pictures
US7394916B2 (en) * 2003-02-10 2008-07-01 Activeye, Inc. Linking tracked objects that undergo temporary occlusion
US20040233983A1 (en) * 2003-05-20 2004-11-25 Marconi Communications, Inc. Security system
US20040252193A1 (en) * 2003-06-12 2004-12-16 Higgins Bruce E. Automated traffic violation monitoring and reporting system with combined video and still-image data
US20040252194A1 (en) * 2003-06-16 2004-12-16 Yung-Ting Lin Linking zones for object tracking and camera handoff
US20050025313A1 (en) * 2003-06-19 2005-02-03 Wachtel Robert A. Digital imaging system for creating a wide-angle image from multiple narrow angle images
US20090051682A1 (en) * 2003-08-15 2009-02-26 Werner Gerhard Lonsing Method and apparatus for producing composite images which contain virtual objects
US7750926B2 (en) * 2003-08-15 2010-07-06 Werner Gerhard Lonsing Method and apparatus for producing composite images which contain virtual objects
US20050047646A1 (en) * 2003-08-27 2005-03-03 Nebojsa Jojic System and method for fast on-line learning of transformed hidden Markov models
US20050073585A1 (en) * 2003-09-19 2005-04-07 Alphatech, Inc. Tracking systems and methods
US20050073685A1 (en) * 2003-10-03 2005-04-07 Olympus Corporation Image processing apparatus and method for processing images
US20050185053A1 (en) * 2004-02-23 2005-08-25 Berkey Thomas F. Motion targeting system and method
US7231065B2 (en) * 2004-03-15 2007-06-12 Embarcadero Systems Corporation Method and apparatus for controlling cameras and performing optical character recognition of container code and chassis code
US20050212918A1 (en) * 2004-03-25 2005-09-29 Bill Serra Monitoring system and method
US20070279494A1 (en) * 2004-04-16 2007-12-06 Aman James A Automatic Event Videoing, Tracking And Content Generation
US20070247525A1 (en) * 2004-06-01 2007-10-25 L-3 Comminications Corporation Video Flashlight/Vision Alert
US20050275721A1 (en) * 2004-06-14 2005-12-15 Yusuke Ishii Monitor system for monitoring suspicious object
US20060007242A1 (en) * 2004-07-08 2006-01-12 Microsoft Corporation Matching digital information flow to a human perception system
US20060020990A1 (en) * 2004-07-22 2006-01-26 Mceneaney Ian P System and method for selectively providing video of travel destinations
US20060028552A1 (en) * 2004-07-28 2006-02-09 Manoj Aggarwal Method and apparatus for stereo, multi-camera tracking and RF and video track fusion
US20070258009A1 (en) * 2004-09-30 2007-11-08 Pioneer Corporation Image Processing Device, Image Processing Method, and Image Processing Program
US20060117356A1 (en) * 2004-12-01 2006-06-01 Microsoft Corporation Interactive montages of sprites for indexing and summarizing video
US20070183770A1 (en) * 2004-12-21 2007-08-09 Katsuji Aoki Camera terminal and imaging zone adjusting apparatus
US20060149829A1 (en) * 2004-12-31 2006-07-06 Ta-Chiun Kuan Monitor system
US20060171453A1 (en) * 2005-01-04 2006-08-03 Rohlfing Thomas R Video surveillance system
US7620266B2 (en) * 2005-01-20 2009-11-17 International Business Machines Corporation Robust and efficient foreground analysis for real-time video surveillance
US20060174302A1 (en) * 2005-02-01 2006-08-03 Bryan Mattern Automated remote monitoring system for construction sites
US20070065002A1 (en) * 2005-02-18 2007-03-22 Laurence Marzell Adaptive 3D image modelling system and apparatus and method therefor
US20060192839A1 (en) * 2005-02-25 2006-08-31 Samsung Electronics Co., Ltd. Method and apparatus for measuring optical density of image printed on medium
US20060197839A1 (en) * 2005-03-07 2006-09-07 Senior Andrew W Automatic multiscale image acquisition from a steerable camera
US20060215765A1 (en) * 2005-03-25 2006-09-28 Cherng-Daw Hwang Split screen video in a multimedia communication system
US20060227138A1 (en) * 2005-04-05 2006-10-12 Nissan Motor Co., Ltd. Image processing device and method
US20060252554A1 (en) * 2005-05-03 2006-11-09 Tangam Technologies Inc. Gaming object position analysis and tracking
US20060252521A1 (en) * 2005-05-03 2006-11-09 Tangam Technologies Inc. Table game tracking
US20070024706A1 (en) * 2005-08-01 2007-02-01 Brannon Robert H Jr Systems and methods for providing high-resolution regions-of-interest
US20070058717A1 (en) * 2005-09-09 2007-03-15 Objectvideo, Inc. Enhanced processing for scanning video
US20070098001A1 (en) * 2005-10-04 2007-05-03 Mammen Thomas PCI express to PCI express based low latency interconnect scheme for clustering systems
US20070104328A1 (en) * 2005-11-04 2007-05-10 Sunplus Technology Co., Ltd. Image signal processing device
US7817207B2 (en) * 2005-11-04 2010-10-19 Sunplus Technology Co., Ltd. Image signal processing device
US20080060034A1 (en) * 2006-02-13 2008-03-06 Geoffrey Egnal System and method to combine multiple video streams
US20070250898A1 (en) * 2006-03-28 2007-10-25 Object Video, Inc. Automatic extraction of secondary video streams
US20080019566A1 (en) * 2006-07-21 2008-01-24 Wolfgang Niem Image-processing device, surveillance system, method for establishing a scene reference image, and computer program
US20080036864A1 (en) * 2006-08-09 2008-02-14 Mccubbrey David System and method for capturing and transmitting image data streams
US20080074494A1 (en) * 2006-09-26 2008-03-27 Harris Corporation Video Surveillance System Providing Tracking of a Moving Object in a Geospatial Model and Related Methods
US20080096372A1 (en) * 2006-10-23 2008-04-24 Interuniversitair Microelektronica Centrum (Imec) Patterning of doped poly-silicon gates
US20080133767A1 (en) * 2006-11-22 2008-06-05 Metis Enterprise Technologies Llc Real-time multicast peer-to-peer video streaming platform
US20080151049A1 (en) * 2006-12-14 2008-06-26 Mccubbrey David L Gaming surveillance system and method of extracting metadata from multiple synchronized cameras
US20080297587A1 (en) * 2007-05-31 2008-12-04 Kurtz Andrew F Multi-camera residential communication system
US8063929B2 (en) * 2007-05-31 2011-11-22 Eastman Kodak Company Managing scene transitions for video communication
US20090086023A1 (en) * 2007-07-18 2009-04-02 Mccubbrey David L Sensor system including a configuration of the sensor as a virtual sensor device
US20110115909A1 (en) * 2009-11-13 2011-05-19 Sternberg Stanley R Method for tracking an object through an environment across multiple cameras

Cited By (106)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080148227A1 (en) * 2002-05-17 2008-06-19 Mccubbrey David L Method of partitioning an algorithm between hardware and software
US8230374B2 (en) 2002-05-17 2012-07-24 Pixel Velocity, Inc. Method of partitioning an algorithm between hardware and software
US20080036864A1 (en) * 2006-08-09 2008-02-14 Mccubbrey David System and method for capturing and transmitting image data streams
US20080151049A1 (en) * 2006-12-14 2008-06-26 Mccubbrey David L Gaming surveillance system and method of extracting metadata from multiple synchronized cameras
US8587661B2 (en) * 2007-02-21 2013-11-19 Pixel Velocity, Inc. Scalable system for wide area surveillance
US8797409B1 (en) 2007-02-28 2014-08-05 Fluke Corporation Thermal imaging device, method, and system providing local and remote displays
US8300981B1 (en) * 2007-02-28 2012-10-30 Fluke Corporation System and method for analyzing multiple thermal images
US8339457B1 (en) 2007-02-28 2012-12-25 Fluke Corporation Systems and methods for time-shared memory access in a thermal imaging device
US8350908B2 (en) 2007-05-22 2013-01-08 Vidsys, Inc. Tracking people and objects using multiple live and recorded surveillance camera video feeds
US20080292140A1 (en) * 2007-05-22 2008-11-27 Stephen Jeffrey Morris Tracking people and objects using multiple live and recorded surveillance camera video feeds
US20090086023A1 (en) * 2007-07-18 2009-04-02 Mccubbrey David L Sensor system including a configuration of the sensor as a virtual sensor device
US20100046810A1 (en) * 2008-08-20 2010-02-25 Fujitsu Limited Fingerprint image acquiring device, fingerprint authenticating apparatus, fingerprint image acquiring method, and fingerprint authenticating method
US8224043B2 (en) * 2008-08-20 2012-07-17 Fujitsu Limited Fingerprint image acquiring device, fingerprint authenticating apparatus, fingerprint image acquiring method, and fingerprint authenticating method
US9071626B2 (en) 2008-10-03 2015-06-30 Vidsys, Inc. Method and apparatus for surveillance system peering
US8525880B2 (en) * 2008-11-21 2013-09-03 Robert Bosch Gmbh Security system including modular ring housing
US20140002648A1 (en) * 2008-11-21 2014-01-02 Bosch Security Systems, Inc. Security system including modular ring housing
US9485477B2 (en) * 2008-11-21 2016-11-01 Robert Bosch Gmbh Security system including modular ring housing
US9578291B2 (en) * 2008-11-21 2017-02-21 Robert Bosch Gmbh Security system including modular ring housing
US20130342982A1 (en) * 2008-11-21 2013-12-26 Bosch Security Systems, Inc. Security system including modular ring housing
US20100128446A1 (en) * 2008-11-21 2010-05-27 Bosch Security Systems, Inc. Security system including modular ring housing
US20130301881A1 (en) * 2008-11-27 2013-11-14 Sony Corporation Image processing apparatus, image processing method, and program
US8494217B2 (en) * 2008-11-27 2013-07-23 Sony Corporation Image processing apparatus, image processing method, and program
US20100128928A1 (en) * 2008-11-27 2010-05-27 Sony Corporation Image processing apparatus, image processing method, and program
US20120265917A1 (en) * 2009-11-09 2012-10-18 Imec Data transferring device
US20110115909A1 (en) * 2009-11-13 2011-05-19 Sternberg Stanley R Method for tracking an object through an environment across multiple cameras
US9955209B2 (en) 2010-04-14 2018-04-24 Alcatel-Lucent Usa Inc. Immersive viewer, a method of providing scenes on a display and an immersive viewing system
US9294716B2 (en) 2010-04-30 2016-03-22 Alcatel Lucent Method and system for controlling an imaging system
US9183560B2 (en) 2010-05-28 2015-11-10 Daniel H. Abelow Reality alternate
US11222298B2 (en) 2010-05-28 2022-01-11 Daniel H. Abelow User-controlled digital environment across devices, places, and times with continuous, variable digital boundaries
US8588533B2 (en) 2010-07-28 2013-11-19 International Business Machines Corporation Semantic parsing of objects in video
US9330312B2 (en) 2010-07-28 2016-05-03 International Business Machines Corporation Multispectral detection of personal attributes for video surveillance
US8774522B2 (en) 2010-07-28 2014-07-08 International Business Machines Corporation Semantic parsing of objects in video
US10424342B2 (en) 2010-07-28 2019-09-24 International Business Machines Corporation Facilitating people search in video surveillance
US9245186B2 (en) 2010-07-28 2016-01-26 International Business Machines Corporation Semantic parsing of objects in video
US9679201B2 (en) 2010-07-28 2017-06-13 International Business Machines Corporation Semantic parsing of objects in video
US8515127B2 (en) 2010-07-28 2013-08-20 International Business Machines Corporation Multispectral detection of personal attributes for video surveillance
US9002117B2 (en) 2010-07-28 2015-04-07 International Business Machines Corporation Semantic parsing of objects in video
US9134399B2 (en) 2010-07-28 2015-09-15 International Business Machines Corporation Attribute-based person tracking across multiple cameras
US8532390B2 (en) 2010-07-28 2013-09-10 International Business Machines Corporation Semantic parsing of objects in video
US9274204B2 (en) * 2010-08-16 2016-03-01 Korea Research Institute Of Standards And Science Camera tracing and surveillance system and method for security using thermal image coordinate
US20130135468A1 (en) * 2010-08-16 2013-05-30 Korea Research Institute Of Standards And Science Camera tracing and surveillance system and method for security using thermal image coordinate
US8754925B2 (en) 2010-09-30 2014-06-17 Alcatel Lucent Audio source locator and tracker, a method of directing a camera to view an audio source and a video conferencing terminal
US20120083314A1 (en) * 2010-09-30 2012-04-05 Ng Hock M Multimedia Telecommunication Apparatus With Motion Tracking
US20120134282A1 (en) * 2010-11-30 2012-05-31 Nokia Corporation Method and apparatus for selecting devices to form a community
US8278779B2 (en) 2011-02-07 2012-10-02 General Electric Company System and method for providing redundant power to a device
US9008487B2 (en) 2011-12-06 2015-04-14 Alcatel Lucent Spatial bookmarking
US20130250134A1 (en) * 2012-02-19 2013-09-26 Jack J. McCauley System and Methods for Wireless Remote Control over Cameras
US9749515B2 (en) * 2012-02-19 2017-08-29 Jack J. McCauley System and methods for wireless remote control over cameras with audio processing to generate a refined audio signal
US8929596B2 (en) 2012-06-04 2015-01-06 International Business Machines Corporation Surveillance including a modified video data stream
US8917909B2 (en) 2012-06-04 2014-12-23 International Business Machines Corporation Surveillance including a modified video data stream
TWI510082B (en) * 2012-06-06 2015-11-21 Etron Technology Inc Image capturing method for image rcognition and system thereof
US9373035B2 (en) * 2012-06-06 2016-06-21 Eys3D Microelectronics, Co. Image capturing method for image recognition and system thereof
US20130329947A1 (en) * 2012-06-06 2013-12-12 Etron Technology, Inc. Image capturing method for image recognition and system thereof
US9087386B2 (en) 2012-11-30 2015-07-21 Vidsys, Inc. Tracking people and objects using multiple live and recorded surveillance camera video feeds
US9706264B2 (en) * 2013-02-06 2017-07-11 Robert Bosch Gmbh Multiple field-of-view video streaming
US20140218612A1 (en) * 2013-02-06 2014-08-07 Robert Bosch Gmbh Multiple field-of-view video streaming
US10142533B2 (en) 2015-03-27 2018-11-27 Intel Corporation Technologies for controlling user access to image sensors of a camera device
EP3275183A4 (en) * 2015-03-27 2018-10-03 Intel Corporation Technologies for controlling user access to image sensors of a camera device
US10332089B1 (en) * 2015-03-31 2019-06-25 Amazon Technologies, Inc. Data synchronization system
US10602060B2 (en) * 2015-07-21 2020-03-24 Hangzhou Hikvision Digital Technology Co., Ltd. Camera and integrated circuit board
US20180220069A1 (en) * 2015-07-21 2018-08-02 Hangzhou Hikvision Digital Technology Co., Ltd. Camera and integrated circuit board
US10491810B2 (en) * 2016-02-29 2019-11-26 Nokia Technologies Oy Adaptive control of image capture parameters in virtual reality cameras
US20190052800A1 (en) * 2016-02-29 2019-02-14 Nokia Technologies Oy Adaptive control of image capture parameters in virtual reality cameras
US10395340B2 (en) * 2016-03-10 2019-08-27 Dell Products, Lp System and method of generating 3D infrared camera stitching for a thermal mapping
US20170264882A1 (en) * 2016-03-10 2017-09-14 Dell Products, Lp System and method of generating a temperature map
US10805507B2 (en) * 2016-12-21 2020-10-13 Shanghai Xiaoyi Technology Co., Ltd. Method and system for configuring cameras to capture images
US20180176425A1 (en) * 2016-12-21 2018-06-21 Shanghai Xiaoyi Technology Co., Ltd. Method and system for configuring cameras to capture images
US11551079B2 (en) 2017-03-01 2023-01-10 Standard Cognition, Corp. Generating labeled training images for use in training a computational neural network for object or action recognition
US11790682B2 (en) 2017-03-10 2023-10-17 Standard Cognition, Corp. Image analysis using neural networks for pose and action identification
US10094662B1 (en) 2017-03-28 2018-10-09 Trimble Inc. Three-dimension position and heading solution
US10406645B2 (en) 2017-05-24 2019-09-10 Trimble Inc. Calibration approach for camera placement
US10341618B2 (en) 2017-05-24 2019-07-02 Trimble Inc. Infrastructure positioning camera system
US10300573B2 (en) 2017-05-24 2019-05-28 Trimble Inc. Measurement, layout, marking, firestop stick
US10646975B2 (en) 2017-05-24 2020-05-12 Trimble Inc. Measurement, layout, marking, firestop stick
US11402271B2 (en) * 2017-06-03 2022-08-02 Teledyne Flir, Llc Extensible architecture for surveillance and targeting imaging systems and methods
US11270260B2 (en) 2017-08-07 2022-03-08 Standard Cognition Corp. Systems and methods for deep learning-based shopper tracking
US11200692B2 (en) 2017-08-07 2021-12-14 Standard Cognition, Corp Systems and methods to check-in shoppers in a cashier-less store
US10474988B2 (en) * 2017-08-07 2019-11-12 Standard Cognition, Corp. Predicting inventory events using foreground/background processing
US10650545B2 (en) 2017-08-07 2020-05-12 Standard Cognition, Corp. Systems and methods to check-in shoppers in a cashier-less store
US10474992B2 (en) 2017-08-07 2019-11-12 Standard Cognition, Corp. Machine learning-based subject tracking
US10474991B2 (en) 2017-08-07 2019-11-12 Standard Cognition, Corp. Deep learning-based store realograms
US10853965B2 (en) 2017-08-07 2020-12-01 Standard Cognition, Corp Directional impression analysis using deep learning
US20190156277A1 (en) * 2017-08-07 2019-05-23 Standard Cognition, Corp Realtime inventory location management using deep learning
US10474993B2 (en) 2017-08-07 2019-11-12 Standard Cognition, Corp. Systems and methods for deep learning-based notifications
US11023850B2 (en) * 2017-08-07 2021-06-01 Standard Cognition, Corp. Realtime inventory location management using deep learning
US11195146B2 (en) 2017-08-07 2021-12-07 Standard Cognition, Corp. Systems and methods for deep learning-based shopper tracking
US11538186B2 (en) 2017-08-07 2022-12-27 Standard Cognition, Corp. Systems and methods to check-in shoppers in a cashier-less store
US10445694B2 (en) 2017-08-07 2019-10-15 Standard Cognition, Corp. Realtime inventory tracking using deep learning
US11544866B2 (en) 2017-08-07 2023-01-03 Standard Cognition, Corp Directional impression analysis using deep learning
US11232687B2 (en) 2017-08-07 2022-01-25 Standard Cognition, Corp Deep learning-based shopper statuses in a cashier-less store
US11250376B2 (en) 2017-08-07 2022-02-15 Standard Cognition, Corp Product correlation analysis using deep learning
US11810317B2 (en) 2017-08-07 2023-11-07 Standard Cognition, Corp. Systems and methods to check-in shoppers in a cashier-less store
US11295270B2 (en) 2017-08-07 2022-04-05 Standard Cognition, Corp. Deep learning-based store realograms
US20190043003A1 (en) * 2017-08-07 2019-02-07 Standard Cognition, Corp Predicting inventory events using foreground/background processing
US10347008B2 (en) 2017-08-14 2019-07-09 Trimble Inc. Self positioning camera system to 3D CAD/BIM model
US10339670B2 (en) 2017-08-29 2019-07-02 Trimble Inc. 3D tool tracking and positioning using cameras
US11232575B2 (en) * 2019-04-18 2022-01-25 Standard Cognition, Corp Systems and methods for deep learning-based subject persistence
US11948313B2 (en) 2019-04-18 2024-04-02 Standard Cognition, Corp Systems and methods of implementing multiple trained inference engines to identify and track subjects over multiple identification intervals
US10997747B2 (en) 2019-05-09 2021-05-04 Trimble Inc. Target positioning with bundle adjustment
US11610142B2 (en) * 2019-05-28 2023-03-21 Ati Technologies Ulc Safety monitor for image misclassification
US11002541B2 (en) 2019-07-23 2021-05-11 Trimble Inc. Target positioning with electronic distance measuring and bundle adjustment
US11641288B2 (en) * 2019-09-19 2023-05-02 Telefonaktiebolaget Lm Ericsson (Publ) Low latency small message communication channel using remote write
US20220417055A1 (en) * 2019-09-19 2022-12-29 Telefonaktiebolaget Lm Ericsson (Publ) Low latency small message communication channel using remote write
US11361468B2 (en) 2020-06-26 2022-06-14 Standard Cognition, Corp. Systems and methods for automated recalibration of sensors for autonomous checkout
US11818508B2 (en) 2020-06-26 2023-11-14 Standard Cognition, Corp. Systems and methods for automated design of camera placement and cameras arrangements for autonomous checkout
US11303853B2 (en) 2020-06-26 2022-04-12 Standard Cognition, Corp. Systems and methods for automated design of camera placement and cameras arrangements for autonomous checkout

Also Published As

Publication number Publication date
GB2459602A (en) 2009-11-04
JP2010519860A (en) 2010-06-03
IL200553A0 (en) 2010-05-17
US8587661B2 (en) 2013-11-19
GB0914684D0 (en) 2009-09-30
GB2459602B (en) 2011-09-21
WO2008103850A2 (en) 2008-08-28
WO2008103850A3 (en) 2008-10-16

Similar Documents

Publication Publication Date Title
US8587661B2 (en) Scalable system for wide area surveillance
US10122998B2 (en) Real time sensor and method for synchronizing real time sensor data streams
JP6894687B2 (en) Image processing system, image processing device, control method, and program
Rowe et al. Firefly mosaic: A vision-enabled wireless sensor networking system
Yang et al. TEEVE: The next generation architecture for tele-immersive environments
KR101208427B1 (en) Multiple Camera control and image storing apparatus and method for synchronized multiple image acquisition
CN101505434A (en) High resolution intelligent network camera array system having global synchronization function
KR100744668B1 (en) Multiple person surveillance system using networked cameras
US20170339329A1 (en) Intelligence Interface for Interchangable Sensors
US20200162719A1 (en) Systems, methods and apparatuses for stereo vision
US10951804B2 (en) Photographing synchronization method and apparatus
CN102932676B (en) Self-adaptive bandwidth transmitting and playing method based on audio and video frequency synchronization
KR101821145B1 (en) Video live streaming system
CN111970524B (en) Control method, device, system, equipment and medium for interactive live broadcast and microphone connection
CN108616724A (en) A kind of radio image collecting transmission platform based on FPGA and ARM structures
CN106210503B (en) Camera shutter synchronous control method and equipment in vehicle-mounted ethernet communication network
CN107800946A (en) A kind of live broadcasting method and system
JP2015508584A (en) Method for 3D reconstruction of scenes that rely on asynchronous sensors
Kandhalu et al. DSPcam: A camera sensor system for surveillance networks
JP6821398B2 (en) Image processing equipment, image processing methods and programs
US11398076B2 (en) Three dimensional media streaming and broadcasting system and method
CN108737809A (en) A kind of remote synchronization image-pickup method
Ibraheem et al. A resource-efficient multi-camera gige vision ip core for embedded vision processing platforms
US6684402B1 (en) Control methods and apparatus for coupling multiple image acquisition devices to a digital data processor
WO2017049597A1 (en) System and method for video broadcasting

Legal Events

Date Code Title Description
AS Assignment

Owner name: PIXEL VELOCITY, INC., MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MCCUBBREY, DAVID L;REEL/FRAME:031390/0081

Effective date: 20080501

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.)

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20171119