US9591272B2 - Optimal camera selection in array of monitoring cameras - Google Patents
Optimal camera selection in array of monitoring cameras Download PDFInfo
- Publication number
- US9591272B2 US9591272B2 US14/115,093 US201314115093A US9591272B2 US 9591272 B2 US9591272 B2 US 9591272B2 US 201314115093 A US201314115093 A US 201314115093A US 9591272 B2 US9591272 B2 US 9591272B2
- Authority
- US
- United States
- Prior art keywords
- resolution
- camera
- cameras
- coverage
- cost
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000012544 monitoring process Methods 0.000 title claims abstract description 49
- 238000000034 method Methods 0.000 claims abstract description 89
- 230000006870 function Effects 0.000 claims description 42
- 239000011159 matrix material Substances 0.000 claims description 27
- 239000013598 vector Substances 0.000 claims description 19
- 238000005457 optimization Methods 0.000 claims description 17
- 230000008569 process Effects 0.000 claims description 12
- 230000004044 response Effects 0.000 claims description 12
- 230000007423 decrease Effects 0.000 claims description 6
- 230000009467 reduction Effects 0.000 claims description 5
- 238000001514 detection method Methods 0.000 claims 4
- 238000005516 engineering process Methods 0.000 abstract description 6
- 230000007704 transition Effects 0.000 description 37
- 238000010586 diagram Methods 0.000 description 27
- 238000004891 communication Methods 0.000 description 24
- 230000004083 survival effect Effects 0.000 description 13
- 238000013459 approach Methods 0.000 description 12
- 238000004590 computer program Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 238000005286 illumination Methods 0.000 description 5
- 239000007787 solid Substances 0.000 description 4
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 230000009471 action Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 230000001815 facial effect Effects 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000012804 iterative process Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 241000614201 Adenocaulon bicolor Species 0.000 description 1
- AZFKQCNGMSSWDS-UHFFFAOYSA-N MCPA-thioethyl Chemical compound CCSC(=O)COC1=CC=C(Cl)C=C1C AZFKQCNGMSSWDS-UHFFFAOYSA-N 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000003153 chemical reaction reagent Substances 0.000 description 1
- 150000001875 compounds Chemical class 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 201000002266 mite infestation Diseases 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 239000003607 modifier Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007723 transport mechanism Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/62—Control of parameters via user interfaces
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19639—Details of the system layout
- G08B13/19641—Multiple cameras having overlapping views on a single scene
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- H04N5/247—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
Definitions
- a surveillance operator may be able select suitable cameras and resolutions for an area such that the number of cameras and the error between the assigned and desired resolutions are minimized simultaneously.
- the task may become too complicated to be resolved manually.
- the present disclosure generally describes methods, apparatus, systems, devices, and/or computer program products related to automatically optimizing the efficiency of camera placement, numbers, and resolution for a multi-camera monitoring and surveillance application.
- Example methods may include determining a maximum resolution matrix, V, where each element, v i,j , of the V represents a maximum resolution with which a camera c i is capable to monitor a point p j in the multi-camera environment; receiving a desired resolution vector, Res des , where each element of the Res des represents a desired resolution for each point; and evaluating the elements of the V in view of the Res des to determine an optimal camera and resolution selection taking into consideration a cost function, where the cost function includes at least an error in a resolution assigned to each point.
- a computing device operable to automatically optimize an efficiency of camera placement, numbers, and resolution in a multi-camera monitoring environment.
- the computing device may include a memory configured to store instructions; an input device configured to receive a desired resolution vector, Res des , where each element of the Res des represents a desired resolution for each point in the multi-camera environment; and a processor.
- the processor may be configured to determine a maximum resolution matrix, V, where each element, v i,j , of the V represents a maximum resolution with which a camera c i is capable to monitor a point p j in the multi-camera environment; and evaluate the elements of the V in view of the Res des to determine an optimal camera and resolution selection taking into consideration a cost function, where the cost function includes at least an error in a resolution assigned to each point.
- An example method may include determining a plurality of resolutions associated with a plurality of cameras defined for intervals along a linear axis; receiving information associated with points on the intervals and desired resolutions for the points; forming a combinatorial state trellis, where each level represents a point according to a linear order of the points and possible combinations of camera resolutions covering the point are listed as states on a corresponding level; and evaluating optimal paths through the levels while obeying resolution constraints in each path that is traversed in the trellis until a survival path is determined.
- a computing device for optimal camera selection in array of cameras for monitoring and surveillance applications.
- the computing device may include a memory configured to store instructions and a processor.
- the processor may be configured to determine a plurality of resolutions associated with a plurality of cameras defined for intervals along a linear axis; receive information associated with points on the intervals and desired resolutions for the points; form a combinatorial state trellis, where each level represents a point according to a linear order of the points and possible combinations of camera resolutions covering the point are listed as states on a corresponding level; and evaluate optimal paths through the levels while obeying resolution constraints in each path that is traversed in the trellis until a survival path is determined.
- a method for optimal light subset selection in a lighting array that achieves a desired intensity for an area of illumination.
- An example method may include determining a plurality of lighting intensities associated with a plurality of lights defined for intervals along a linear axis; receiving information associated with points on the intervals and desired lighting intensities for the points; forming a combinatorial state trellis, wherein each level represents a point according to a linear order of the points and possible combinations of lighting intensities covering the point are listed as states on a corresponding level; and evaluating optimal paths through the levels while obeying lighting intensity constraints in each path that is traversed in the trellis until a survival path is determined.
- the processor may be configured to determine a plurality of lighting intensities associated with a plurality of lights defined for intervals along a linear axis; receive information associated with points on the intervals and desired lighting intensities for the points; form a combinatorial state trellis, wherein each level represents a point according to a linear order of the points and possible combinations of lighting intensities covering the point are listed as states on a corresponding level; and evaluate optimal paths through the levels while obeying lighting intensity constraints in each path that is traversed in the trellis until a survival path is determined.
- a computer readable storage medium with instructions stored thereon for executing the above methods at one or more processors for optimizing an efficiency of camera placement, numbers, and resolution in a multi-camera monitoring environment may also be described.
- FIG. 1 illustrates an example two-dimensional system of a number of points and multi-resolution cameras
- FIG. 2 illustrates a block diagram of an example system employing the greedy method for optimal camera selection in an array of cameras for monitoring and surveillance applications;
- FIG. 3 illustrates an example one-dimensional scenario where people in a subway station are to be covered with cameras having two different resolutions
- FIG. 4 illustrates an interval covering computation with a solid line segment that is divided into four sub-intervals
- FIG. 5 illustrates an example one-dimensional single resolution scenario of six cameras with different coverage areas
- FIG. 6 illustrates another example one-dimensional multi-resolution camera configuration with three cameras, where each camera has two resolutions and four points to be covered;
- FIG. 7 illustrates a further example one-dimensional multi-resolution camera configuration with four cameras, where each camera has two resolutions and three points to be covered;
- FIG. 8 illustrates a block diagram of an example system employing a combinatorial state Viterbi technique for optimal camera selection in an array of cameras for monitoring and surveillance applications;
- FIG. 9 illustrates a trellis of the combinatorial states trellis technique when applied to the example scenario of FIG. 7 ;
- FIG. 10 illustrates a general purpose computing device, which may be used to manage optimal camera selection in an array of cameras for monitoring and surveillance applications;
- FIG. 11 is a flow diagram illustrating an example method that may be performed by a computing device such as the computing device in FIG. 10 ;
- FIG. 12 illustrates a block diagram of an example computer program product, all arranged in accordance with a least some embodiments described herein.
- This disclosure is generally drawn, inter alia, to methods, apparatus, systems, devices, and/or computer program products related to automatic optimization of efficiency of camera placement, numbers, and resolution for a multi-camera monitoring and surveillance application.
- technologies are generally provided for automatically optimizing an efficiency of camera placement, numbers, and resolution in multi-camera monitoring and surveillance applications.
- a fraction of a total area may be monitored at a higher resolution than the rest.
- Employing techniques such as combinatorial state Viterbi technique or combinatorial state trellis technique, a minimum number of cameras that provide the coverage at the needed resolution may be selected.
- a number of points may be covered with at least a predefined number of cameras. For example, a subject of interest may be tracked in a public area, where specific camera(s) may be used to image the subject's face at a higher resolution than the background.
- FIG. 1 illustrates an example tow-dimensional system of a number of points and multi-resolutions cameras, arranged in accordance with at least some embodiments described herein.
- Two-dimensional coverage in area surveillance typically assumes that the objects in the coverage area are seen at a single resolution. This may not always be the case, however, and multiple resolutions (for example, for facial recognition purposes) may be desirable.
- a system according to some examples treats coverage of two-dimensional surveillance with multiple resolution selection as an NP-complete configuration employs heuristic techniques to select cameras in an array of cameras for monitoring and surveillance applications. In some example scenarios, the two-dimensional configuration may be converted to an equivalent one-dimensional configuration without loss of optimality.
- monitoring/surveillance system may include a number of cameras such as cameras C1 through C8.
- the cameras C1 through C8 may be positioned such that a number of points p1 through p4 are covered at specific resolutions, among other things.
- Each camera may have its coverage area such as a coverage area 102 for the camera C1, a coverage area 102 for the camera C1, a coverage area 104 for the camera C2, a coverage area 106 for the camera C3, a coverage area 108 for the camera C4, a coverage area 110 for the camera C5, a coverage area 112 for the camera C6, a coverage area 114 for the camera C7, and coverage area 116 for the camera C8.
- One or more of the cameras in a system may be settable to different resolutions, where each resolution may represent a level of zoom. Thus, higher resolution may provide a more magnified image.
- a higher resolution coverage area of a camera may be smaller than the coverage area for a lower resolution as depicted by the coverage areas for different resolutions (1 through 3) of the camera C1.
- the cameras may be configured to cover a number of points with specific resolutions. A number of points and a number of cameras are referred to as NoP and NoC, respectively, herein.
- the assigned resolutions to cameras may be represented with a vector A, where each element Xi corresponds to a resolution of a camera i.
- a length of the vector X is, thus, NoC.
- X [X 1 X 2 . . . X NoC ] [1]
- n R a maximum resolution value
- each vector element may be considered as Xi ⁇ 0, 1, 2, . . . , n R ⁇ [2] in a general case, where points and cameras may be distributed over a two-dimensional area at any location.
- Each point may be desired to be covered with a specific resolution.
- the decision involves which cameras to turn on and what resolution to assign to each camera so that the desired resolution is approximately provided for all of the points.
- the number of “ON” cameras and an error in assigned resolution to each point may be minimized simultaneously.
- a system according to embodiments may model this configuration approach as a discrete optimization.
- V a maximum resolution matrix
- V may be an NoC by NoP matrix, where each element of V, v i,j , may represent the maximum resolution with which camera i can see point j. If it can be determined that a point can be covered by a camera with at most resolution k, the point may also be covered by the same camera with lower resolutions and cannot be seen with higher resolutions. Thus, the knowledge of the maximum resolution may include the information about other resolutions as well.
- the maximum resolution matrix, V may be used to define one of the costs as discussed below.
- One of the cost functions may include the error in the resolution assigned to each point.
- the desired resolution of the point j Res desj
- Res assj the assigned resolution of the point j
- Elements of X which are less than or equal to their corresponding element in j th column of V may be considered and the maximum selected. If a resolution higher than the desired resolution is assigned to a point, the error in the resolution assignment may be considered to be zero.
- the camera selection may be modeled as a discrete optimization.
- the number of “ON” cameras cost and the resolution error cost may be linearly combined with a parameter ⁇ to determine a total cost.
- ⁇ is a factor which determines an importance of one part of the cost function over another Part. Subsequently, the total cost may be minimized using:
- ⁇ may have a substantial effect on the solution. For example, if ⁇ is selected large enough, the above described technique may tend to minimize the resolution cost approximately regardless of the first cost. If one cost is not preferred to another, both costs may be normalized to their maximum value and ⁇ selected to be 1. If the number of cameras is less than the number of points, then the maximum value of I o (X) may be NoC. IF the number of cameras is not less than the number of points (i.e. NoP ⁇ NoC), the approach may be applied with NoP cameras with the maximum value of I o (X) being min(NoC, NoP) in this case. The maximum value of the resolution cost may be obtained when the assigned resolutions are zero:
- FIG. 2 illustrates a block diagram of an example system employing the greedy method for optimal camera selection in an array of cameras for monitoring and surveillance applications, arranged in accordance with at least some embodiments described herein.
- a greedy method may be employed in a resolution-first camera assignment approach.
- Employing the greedy method may provide. Acceptable results with relatively high speed and low complexity, thus avoiding an increased need for computational system resources.
- more weight may be given to the resolution cost rather than the number of ON cameras cost.
- a greedy method may minimize the total cost function iteratively.
- a camera may be selected and a resolution assigned to the camera such that the selected camera and its assigned resolution impose a largest reduction on the cost function among other choices.
- the iterations may continue until no other selections (camera or resolution) decrease the cost function, for example.
- a diagram 200 illustrates how the iterative process may be performed.
- the inputs of the iterative process may include Res des , an array of desired resolutions of target points; ⁇ , a factor which determines the weight of one part of cost function over another part; and V, a matrix 212 which represents the maximum resolution that a camera observes a point.
- the best resolution may be selected through best resolution selection blocks 214 , 216 .
- the best resolution for each camera may be the one that decreases the cost function more than other resolutions.
- the total cost function may be computed for all assignable resolutions.
- the minimum value of computed costs and a corresponding resolution may be provided to an input of a cost analyzer block 218 .
- the cost analyzer block 218 may determine the minimum cost among input costs (Cost 1 , . . . Cost NoC ) and compare the minimum cost with the cost of previous iteration (PreCost).
- the cost analyzer block 218 may terminate the iterations and provide a “Ready” signal. Else, the minimum cost may be fed to the best resolution selection blocks 214 , 216 for the next iteration.
- FIG. 3 illustrates an example one-dimensional scenario where people in a subway station are to be covered with cameras having two different resolutions, arranged in accordance with at least some embodiments described herein.
- a special case of one-dimensional configurations which have specific applications, may be considered.
- the optimal solution may be achieved using combinatorial state Viterbi technique in some embodiments.
- a diagram 300 depicts an example scenario for one-dimensional configurations.
- objects people 333 and 335
- a subway station may be on either side of a train 334 .
- cameras 332 and 336 on either side of the tunnel may monitor the people 333 and 335 on their respective sides.
- An approximation may be made that people 333 and 335 are roughly standing on a line making the optimization a linear one.
- the cameras 332 and 336 may have two different resolutions (e.g., normal coverage are and higher resolution coverage area available via zoom). If a line can be fitted through the points (the people being monitored), for example based on the least squared error, and the points projected on the line such that the V matrix does not change, the optimal selection for the two-dimensional configuration may be determined by computing the one-dimensional configuration without loss of optimality.
- the two-dimensional approach may be simplified to a one-dimensional approach of intervals (parts of the objects on a line) which may be desired to be covered with different resolutions.
- a point covering approach may be employed in the one-dimensional configuration.
- FIG. 4 illustrates an interval covering computation with a solid line segment that is divided into four sub-intervals, arranged in accordance with at least some embodiments described herein.
- each interval may be broken down into a number of sub-intervals and a point may be considered as the representative of each sub-interval. Then the optimization may be performed for the points.
- the interval of the coverage area of i th camera sensor may be represented by [a i , b i ].
- the intervals along the line 440 may be converted to points as follows.
- the beginning (I int ) and the end of interval (u int ) may be considered and the beginnings and ends of the cameras coverage boundaries of any resolutions sorted.
- a diagram 400 depicts cameras C1 and C2 observe an interval under consideration 445 with high resolution coverages 442 and 444 , respectively.
- the interval under consideration 445 is depicted as solid line segment and is segmented to four sub-intervals.
- I int may be I 0 and u int may be I 4 in the example scenario. All points on each sub-interval may be covered with the same cameras and resolutions. Thus, a single point may be considered as a representative of each sub-interval.
- FIG. 5 illustrates an example one-dimensional single resolution scenario of six cameras with different coverage areas, arranged in accordance with at least some embodiments described herein.
- a diagram 500 illustrates an example scenario with a number of cameras (C1 through C6) covering a line 546 , where the cameras may have overlap in their coverage areas.
- no resolution is assigned to the points. Therefore, the cameras' resolutions do not change and the aim is to cover the points with minimum number of cameras.
- a discrete (combinatorial) optimization technique may be applied.
- An assumption may be made that the points may be on a substantially straight line and the coverage of each camera may be defined with an interval.
- the points, p i represent objects or parts of the objects on a real number axis.
- the computation may begin from the most left-hand side point (p 1 ). Starting from the point, p 1 , a camera with the maximum coverage on the real axis may be selected after the p 1 location. This selection may be considered as equivalent to the selection of the interval which contains the point and has the largest b i (the end of the camera coverage interval). By selecting the interval with the point p 1 and the largest b i , any possible optimal selection is not lost because after this selection, any other interval may be selected without any restriction. Thus, optimality is not lost.
- a camera which has the maximum coverage or (in other words the greatest likelihood to cover other points) may be selected.
- the computation may move toward selecting the optimal cameras.
- a determination may be made whether the next point is covered or not. If the next point is covered, the computation may move to the subsequent point. If the next point is not covered, the selection of the first point may be repeated for the next point as well. The selection process may be iteratively repeated until all points are covered.
- the above-described approach may be applied in reverse direction starting at the most right-hand side point (p 5 ) as well selecting cameras with maximum coverage on the real axis from the most right-hand side point toward the left.
- FIG. 6 illustrates another example one-dimensional multi-resolution camera configuration with three cameras, where each camera has two resolutions and four points to be covered, arranged in accordance with at least some embodiments described herein.
- the one-dimensional single resolution scenario is a generalization of the multi-resolution configuration and a special case of the two-dimensional configuration. Therefore, the same cost functions as discussed in the two-dimensional case may be minimized.
- each camera may be set to different resolutions but one at a time in case of the one-dimensional multi-resolution camera configuration.
- different resolutions of single camera may be considered as distinct cameras with different coverage areas or as intervals with different lengths.
- a diagram 600 shows an example configurations with three cameras C1, C2, and C3.
- the cameras C1, C2, and C3 have, each, a low resolution coverage and a high resolution coverage ( 652 , 654 , 656 , respectively).
- Points p1 through p4 to be covered are lined along the line 650 . While the point p1 is in the low resolution coverage area of the camera C1, the points p2, p3, and p4 are in overlapping coverage areas.
- the point p2 is in an overlap area 653 that is covered by the high resolution coverage 652 of the camera C1 and the low resolution coverage of the camera C2.
- the point p3 is in an area covered by the low resolution coverages of the camera C2 and C3.
- the point p4 is in an overlap area 655 covered by the low resolution coverage of the camera C2 and by the high resolution coverage 656 of the camera C3.
- one point may be considered and the others removed. If one of them is covered with a minimum error, the others may be covered with the same error because their situations may be identical and if a camera is assigned to one of them, others may be seen with the same camera and resolution.
- the property for removing some points may be used in this example scenario and the points in each interval may be removed except one.
- the selected point may be the representative of its respective interval and the selection may be arbitrary. As a non-conflicting choice, the middle point of each interval may be selected as its representative, for example.
- the sub-interval length may be multiplied by the desired resolution in order to count a length to the total resolution cost.
- the point covering technique may include selection of a middle point for each sub-interval and the cost may be the desired resolution of that sub-interval multiplied by the length of it.
- FIG. 7 illustrates a further example one-dimensional multi-resolution camera configuration with four cameras, where each camera has two resolutions and three points to be covered, arranged in accordance with a least some embodiments described herein.
- An example scenario is shown in a diagram 700 with four multi-resolution cameras C1 through C4.
- the camera resolutions may vary between 1 and 4 and the three points p1, p2, and p3, may be desired to be covered with resolutions 1, 4, and 1, respectively.
- a combinatorial state trellis technique may be employed as discussed in more detail below in conjunction with FIG. 8 and FIG. 9 .
- the result of the optimization may provide a low resolution coverage 762 of the camera C1 to p1, a high resolution coverage 764 of the camera C2 to p2, and low resolution coverage 766 of the camera C3 to p3.
- the optimization may select resolution zero for the cameras 3 and 4 meaning these cameras are not selected and may be turned off.
- FIG. 8 illustrates a block diagram of an example system employing a combinatorial state Viterbi technique for optimal camera selection in an array of cameras for monitoring and surveillance applications, arranged in accordance with at least some embodiments described herein.
- the example block diagram of the combinatorial state Viterbi technique shown in a diagram 800 includes state generators 882 , 884 , which may take a V matrix 886 as input and generate combinations of cameras and resolutions (i.e. states) with which a target point may be observed.
- the states may be provided from the state generators 882 , 884 to combinatorial state Viterbi implementers 874 , 876 , which with some modifications to the Viterbi algorithm, may select the best path to each state and calculate the cost based on the selected path. While two generators and combinatorial state Viterbi implementers are shown for illustration purposes, a plurality of those blocks may be employed in practical implementations, for example, one for each column for the V matrix 886 .
- the combinatorial state Viterbi implementers 874 , 876 may able receive a Res des vector 872 representing desired resolutions for the different cameras as input.
- a ⁇ parameter defining weighting among the different cameras may also be provided to the combinatorial state Viterbi implementers 874 , 876 .
- a minimum cost path finder block 878 may determine a state among the states of the last branch that results in the minimum cost.
- an exhaustive search which considers all eligible (consistent) cases, computes that cost of each case, and chooses the case with the lowest cost may be employed in selecting optimal camera configuration.
- two different exhaustive search approaches may be employed.
- First, approach (exhaustive search on the points) may consider the points and possible scenarios of camera assignment to the points. After determining the possible (not necessarily eligible) scenarios of cameras and their resolutions, the resolution consistency may be examined and scenarios which violate the consistency may be omitted. Among the remaining scenarios, a scenario with the lowest cost may be selected.
- the camera When a camera is considered for a point, the camera may be set on resolution zero meaning that camera has not been selected (it is off) or set on a resolution that cannot see the point to which the camera is assigned to.
- Another approach may consider the cameras and test possible resolutions for each camera. In this case there may not be a need to check the consistency because each camera is considered once an assigned a resolution (including resolution zero). Depending on the number of points and cameras, the two approaches may differ in time/computational complexity.
- a Viterbi technique may be applied to a well-defined trellis to determine the optimal configuration.
- An optimal low complexity combinational-state trellis that takes resolution constraint into consideration may be used.
- a level may be considered for each point and possible combinations of cameras resolutions covering a point may be listed as states on the level.
- Resolution constraint may be obeyed in each path that is traversed in the trellis.
- transitions may be formed from one level (point) to the next level, while resolution consistency is followed for the common cameras corresponding to the start and end states of each transition. It should be noted that the coverage area of each camera on a real axis is contiguous and includes one segment.
- the resolution consistency may be examined from one level to the next without having to examine previous levels of the path because when the current camera is not seen in the previous level, it means that the current level is the first time that the camera is selected in the examined path.
- the resolution consistency for a camera may also be not needed to be examined in the future states on the path where the camera is not present. Because there is the possibility of making a decision about the future branches to a state and selecting an optimum one (which has the minimum cost) without considering the past and the future of the path, the Viterbi technique may be applied directly to find the survival path in a trellis.
- the total cost of each transition may be defined to be a linear combination of the number of cameras and resolution costs with equal weight, as discussed previously.
- Camera resolution consistency in transitions from one level to the next may allow transition which begins from one resolution of a camera and ends in a different resolution of the same camera. There may be some exceptions in the case of resolution zero (when camera is not selected). Transitions from resolution zero to any other resolutions and vice versa may be allowed subject to the following restrictions.
- Transition from a non-zero resolution to a zero resolution may be allowed if the maximum possible resolution in the next level is less than the resolution in the start of the transition.
- the resolution of the camera corresponding to the start state of the transition may be 2. This means that the camera needs to be set to have a resolution of 2 in the current path. If the camera covers the next level (point) with a resolution of 2 or more, the camera cannot be transitioned to resolution zero because the camera can also see the next point with a resolution of 2 and choice of resolution zero violates the resolution consistency. However, if the maximum available resolution in the next level is 1, a transition may be made from resolution 2 to zero. This scenario may occur when the range of coverage of a camera with a resolution of 2 ends at the beginning of the branch in the previous level and it can only see the next point when set on resolution 1.
- Another constraint may be the dual of the above-discussed constraint and may occur when a transition is attempted from resolution zero to a non-zero one. Such a transition may be performed if the maximum resolution in the previous level is less than the non-zero resolution of the desired state.
- the indication is that the coverage interval of the camera with resolution x has ended before the next level and there may not be a transition to any non-zero resolutions in the future of this path. This means that non-zero to zero and then zero to non-zero transition may not occur anywhere in a path.
- transitions may be monitored for each path and variable (e.g., a “check” variable) may be set to predefined value when such a transition happens and the branch selected as the branch with lowest cost that enters the current state.
- keeping track of such transitions (from non-zero to zero) once for all paths (equivalently once for each camera) may be sufficient and not for every single path. If the check variable is set to the predefined value and the resolution of the current state of the current path is zero, then the transition from non-zero to zero may need to necessarily occur once in the current path. Thus, the check variable may be set to the predefined value if the transition branch from non-zero to zero is selected as the survival path going to the state with resolution zero in the next level.
- Embodiments are not limited to the above-discussed configurations or techniques.
- Other optimization techniques and camera configurations with multi-resolution, multidirectional (e.g. PTZ) cameras may also be selected for optimal configuration using the principles discussed herein.
- the principles discussed here may be applied to lighting systems for generating an optimal light subset in a lighting array that achieves a desired intensity for an area of illumination.
- FIG. 9 illustrates a trellis of the combinatorial state trellis technique when applied to the example scenario of FIG. 7 , arranged in accordance with at least some embodiments described herein.
- a depiction of a trellis for the combinatorial state trellis technique that may be applied to the example scenario of FIG. 7 is shown in a diagram 900 .
- the technique does not decide among different resolutions of a camera in finding the optimal configuration.
- the optimal configuration path is shown with a bold line as the survival path.
- the survival path starts with the options 902 for p1 and C1, where C1 is determined to be ON.
- different configurations of C1 and C2 (options 904 ) are considered for p2 and the combination of C1 having the lowest resolution (1) and C2 having the highest resolution (4) is selected).
- options 906 for all four cameras and p3 are considered leading to the determination that the cameras C2, C3, and C4 are not needed for this point.
- the survival path is (1) ⁇ (1,4) ⁇ (1,0,0,0).
- the same technique may be applied to the k-coverage in multidirectional camera sensors in one-dimensional configurations, a one-dimensional scenario of pan-tilt-zoom (PTZ) cameras, or a one-dimensional case of multi-resolution/multidirectional cameras.
- a multidirectional scenario different combinations of directions may be considered as states on each branch of the trellis.
- Direction 0 for each camera may also be considered representing when the camera is off or a current point cannot be seen with the selected direction of the camera.
- the V matrix may be changed to a direction matrix (D) such that each component d i,j shows the direction in which camera i can see point j.
- D direction matrix
- combinatorial states to two component vectors may be used where the first component may show the direction and the second component may show the maximum resolution with which each camera can see a current point.
- FIG. 10 illustrates a general purpose computing device, which may be used to mange optimal camera selection in an array of cameras for monitoring and surveillance applications, arranged in accordance with at least some embodiments described herein.
- the computing device 1000 may be used as a server, desktop computer, portable computer, smart phone, special purpose computer, or similar device such as a controller at a utility control center or a controller at a micro grid.
- the computing device 1000 may include one or more processors 1004 and a system memory 1006 .
- a memory bus 1008 may be used for communicating between the processor 1004 and the system memory 1006 .
- the basic configuration 1002 is illustrated in FIG. 10 by those components within the inner dashed line.
- the processor 1004 may be any type, including but not limited to a microprocessor ( ⁇ P), a microcontroller ( ⁇ C), a digital signal processor (DSP), or any combination thereof.
- the processor 1004 may include one more levels of caching, such as a cache memory 1012 , one or more processor cores 1014 , and registers 1016 .
- the example processor cores 1014 may (each) include an arithmetic logic unit (ALU), a floating point unit (FPU), a digital signal processing core (DSP Core), or any combination thereof.
- An example memory controller 1018 may also be used with the processor 1004 , or in some implementations the memory controller 1018 may be an internal part of the processor 1004 .
- the system memory 1006 may be of any type including but not limited to volatile memory (such as RAM), non-volatile memory (such as ROM, flash memory, etc.) or any combination thereof.
- the system memory 1006 may include an operating system 1020 , one or more applications 1022 , and program data 1024 .
- the application 1022 may include an optimization module 1026 , which may be an integral part of the application 1022 or a separate application on its own.
- the optimization module 1026 may perform optimal camera selection in an array of cameras for monitoring and surveillance applications, as described herein.
- the program data 1024 may include, among other data, data 1028 related to camera positions, resolutions, or the like, as described herein.
- the computing device 1000 may have additional features or functionality, and additional interfaces to facilitate communications between the basic configuration 1002 and any desired devices and interfaces.
- a bus/interface controller 1030 may be used to facilitate communications between the basic configuration 1002 and one or more data storage devices 1032 via a storage interface bus 1034 .
- the data storage devices 1032 may be one or more removable storage devices 1036 , one or more non-removable storage devices 1038 , or a combination thereof.
- Examples of the removable storage and the non-removable storage devices include magnetic disk devices such as flexible disk drives and hard-disk drives (HDD), optical disk drives such as compact disk (CD) drives or digital versatile disk (DVD) drives, solid state drives (SSD), and tape drives to name a few.
- Example computer storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
- the system memory 1006 , the removable storage devices 1036 and the non-removable storage devices 1038 are examples of computer storage media.
- Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD), solid state drives, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which may be used to store the desired information and which may be accessed by the computing device 1000 . Any such computer storage media may be part of the computing device 1000 .
- the computing device 1000 may also include an interface bus 1040 for facilitating communication from various interface devices (e.g., one or more output devices 1042 , one or more peripheral interfaces 1044 , and one or more communication devices 1066 ) to the basic configuration 1002 via the bus/interface controller 1030 .
- interface devices e.g., one or more output devices 1042 , one or more peripheral interfaces 1044 , and one or more communication devices 1066
- Some of the example output devices 1042 include a graphics processing unit 1048 and an audio processing unit 1050 , which may be configured to communicate to various external devices such as a display or speakers via one or more A/V ports 1052 .
- One or more example peripheral interfaces 1044 may include a serial interface controller 1054 or a parallel interface controller 1056 , which may be configured to communicate with external devices such as input devices (e.g., keyboard, mouse, pen, voice input device, touch input device, etc.) or other peripheral devices (e.g., printer, scanner, etc.) via one or more I/O ports 1058 .
- An example communication device 1066 includes a network controller 1060 , which may be arranged to facilitate communications with one or more other computing devices 1062 over a network communication link via one or more communication ports 1064 .
- the one or more other computing devices 1062 may include servers, camera controller, and comparable devices.
- the network communication link may be one example of a communication media
- communication media may typically be embodied by computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and may include any information delivery media.
- a “modulated data signal” may be a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media may include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), microwave, infrared (IR) and other wireless media.
- RF radio frequency
- IR infrared
- the term computer readable media as used herein may include both storage media and communication media.
- the computing device 1000 may be implemented as a part of a general purpose or Specialized server, mainframe, or similar computer that includes any of the above functions.
- the computing device 1000 may also be implemented as a personal computer including both laptop computer and non-laptop computer configurations.
- Example embodiments may also include methods for maintaining application performances upon transfer between cloud servers. These methods can be implemented in any number of ways, including the structures described herein. One such way may be by machine operations, of devices of the type described in the present disclosure. Another optional way may be for one or more of the individual operations of the methods to be performed in conjunction with one or more human operators performing some of the operations while other operations may be performed by machines. These human operators need not be collocated with each other, but each can be only with a machine that performs a portion of the program. In other embodiments, the human interaction can be automated such as by pre-selected criteria that may be machine automated.
- FIG. 11 is a flow diagram illustrating an example method that may be performed by a computing device such as the computing device in FIG. 10 , arranged in accordance with at least some embodiments described herein.
- Example methods may include one or more operations, functions or actions as illustrated by one or more blocks 1122 , 1124 , 1126 , and/or 1128 .
- the operations described in the blocks 1122 through 1128 may also be stored as computer-executable instructions in a computer-readable medium such as a computer-readable medium 1120 of a computing device 1110 .
- An example process for optimal camera selection in array of cameras for monitoring and surveillance applications may begin with block 1122 , “DETERMINE A MAXIMUM RESOLUTION MATRIX, V. EACH ELEMENT, V I,J , REPRESENTING A MAXIMUM RESOLUTION WITH WHICH A CAMERA C 1 IS CAPABLE TO MONITOR A POINT P J ”, where a maximum resolution matrix such as matrix V ( 212 ) may be determined by a processor of a controller for a monitoring application.
- Block 1122 may be followed by block 1124 , “RECEIVED A DESIRED RESOLUTION VECTOR, RES DES , EACH ELEMENT OF THE RES DES REPRESENTING A DESIRED RESOLUTION FOR EACH POINT P J ”, where desired resolution information may be received, for example, from a user or an automated system such as one that aims to perform facial recognition in a public area.
- Block 1124 may be followed by block 1126 , “EVALUATE THE ELEMENTS OF V IN VIEW OF RES DES TO DETERMINE AN OPTIMAL CAMERA AND RESOLUTION SELECTION TAKING INTO CONSIDERATION A COST FUNCTION”, where the resolutions and cameras may be evaluated through an iterative greedy technique or a combinatorial-state trellis computation (employing Viterbi technique) by a best resolution selection module 216 , for example.
- Block 1126 may be followed by block 1128 , “MINIMIZE COST FUNCTION”, where the cost function may be minimized by a cost analyzer block 218 , for example.
- the evaluation techniques may also be applied in lighting applications, where selected areas may be associated with desired lighting intensities from an array of lights.
- Optimal camera selection in array of cameras for monitoring and surveillance applications may be implemented by similar processes with fewer or additional blocks.
- the blocks may be performed in a different order.
- various blocks may be eliminated.
- various blocks may be divided into additional blocks, or combined together into fewer blocks.
- FIG. 12 illustrates block diagram of an example computer program products, arranged in accordance with at least some embodiments described herein.
- the computer program product 1200 may include a signal bearing medium 1202 that may also include one or more machine readable instructions 1204 that, when executed by, for example, a processor, may provide the functionality described herein.
- an optimization module 1026 executed on the processor 1004 may undertake one or more of the tasks shown in FIG. 12 in response to the instructions 1204 conveyed to the processor 1004 by the signal bearing medium 1202 to perform actions associated with optimal camera selection in array of cameras for monitoring and surveillance applications as described herein.
- Some of those instructions may include, for example, instructions for determining a maximum resolution matrix, V, each element, v i,j , representing a maximum resolution with which a camera c i is capable to monitor a point p j ; receiving a desired resolution vector, Res des , each element of the Res des representing a desired resolution for each point p j ; evaluating the elements of V in view of Res des to determine an optimal camera and resolution selection taking into consideration a cost function; minimizing the cost function according to some embodiments described herein.
- the signal bearing medium 1202 depicted in FIG. 12 may encompass a computer-readable medium 1206 , such as, but not limited to, a hard disk drive, a solid state drive, a Compact Disc (CD), a Digital Versatile Disk (DVD), a digital tape, memory, etc.
- the signal bearing medium 1202 may encompass a recordable medium 1208 , such as, but not limited to, memory, read/write (R/W) CDs, R/W DVDs, etc.
- the signal bearing medium 1202 may encompass a communications medium 1210 , such as, but not limited to, a digital and/or an analog communication medium (e.g., a fiber optic cable, a waveguide, a wired communications link, a wireless communication link, etc.).
- a communications medium 1210 such as, but not limited to, a digital and/or an analog communication medium (e.g., a fiber optic cable, a waveguide, a wired communications link, a wireless communication link, etc.).
- the computer program product 1200 may be conveyed to one or more modules of the processor 1004 of FIG. 10 by an RF signal bearing medium, where the signal bearing medium 1202 is conveyed by the wireless communications medium 1210 (e.g., a wireless communications medium conforming with the IEEE 802.11 standard).
- Example methods may include determining a maximum resolution matrix, V, where each element, v i,j , of the v represents a maximum resolution with which a camera c i is capable to monitor a point p j in the multi-camera environment; receiving a desired resolution vector, Res des , where each element of the Res des represents a desired resolution for each point; and evaluating the elements of the V in view of the Res des to determine an optimal camera and resolution selection taking into consideration a cost function, where the cost function includes at least an error in a resolution assigned to each point.
- the methods may further include evaluating the elements of the V in view of a weighting parameter for the cost function; linearly combining a number of selected cameras' cost and the error in the resolution cost with the weighting parameter to determine a total cost; and/or minimizing the total cost to determine an optimal selection of cameras and resolutions to monitor predefined points in the multi-camera monitoring environment with desired resolutions.
- the methods may also include employing a greedy technique to minimize the total cost iteratively; selecting a camera and assigning a resolution to the selected camera such that the selected camera and the assigned resolution impose a largest reduction on the total cost at each iteration; and continuing the iterations until no other camera or resolution selection decreases the cost function.
- the methods may include receiving the resolutions defined for intervals along a linear axis; and selecting the points to represent the intervals.
- the methods may further include selecting the points as middle points of each interval and/or computing the cost for each interval as a product of a desired resolution for the interval and a length of the interval.
- Each resolution may represent a level of zoom.
- the cameras may include one of: a single resolution camera, a multi-resolution camera, a multidirectional camera, a multi-resolution/multidirectional camera, and pan-tilt-zoom (PTZ) camera.
- the computing device may include a memory configured to store instructions; an input device configured to receive a desired resolution vector, Res des , where each element of the Res des represents a desired resolution for each point in the multi-camera environment; and a processor.
- the processor may be configured to determine a maximum resolution matrix, V, where each element, v i,j , of the V represents a maximum resolution with which a camera c i is a capable to monitor a point p j in the multi-camera environment; and evaluate the elements of the V in view of the Res des to determine an optimal camera and resolution selection taking into consideration a cost function, where the cost function includes at least an error in a resolution assigned to each point.
- the processor may also be configured to evaluate the elements of the V in view of a weighting parameter for the cost function; linearly combine a number of selected cameras' cost and the error in the resolution cost with the weighting parameter to determine a total cost; minimize the total cost to determine an optimal selection of cameras and resolutions to monitor predefined points in the multi-camera monitoring environment with desired resolutions; and/or employ a greedy technique to minimize the total cost iteratively.
- the processor may be further configured to select a camera and assign a resolution to the selected camera such that the selected camera and the assigned resolution impose a largest reduction on the total cost as each iteration; and continue the iterations until no other camera or resolution selection decreases the cost function.
- the processor may also receive the resolutions defined for intervals along a linear axis and select the points to represent the intervals.
- the processor may further select the points as middle points of each interval and compute the cost for each interval as a product of s desired resolution for the interval and a length of the interval.
- Each resolution may represent a level of zoom.
- the cameras may include one of: a single resolution camera, a multi-resolution camera, a multidirectional camera, a multi-resolution/multidirectional camera, and a pan-tilt-zoom (PTZ) camera.
- An example method may include determining a plurality of resolutions associated with a plurality of cameras defined for intervals along a linear axis; receiving information associated with point on the intervals and desired resolutions for the points; forming a combinatorial state trellis, where each level represents a point according to a linear order of the points and possible combinations of camera resolutions covering the point are listed as states on a corresponding level; and evaluating optimal paths through the levels while obeying resolution constraints in each path that is traversed in the trellis until a survival path is determined.
- the method may further include forming transitions from one level to a next level, while following a resolution consistency for common cameras corresponding to a start and an end state of each transition; enabling a transition from a non-zero resolution to a zero resolution in response to a determination that a maximum possible resolution in the next level is less than a resolution at a start of the transition; and/or enabling a transition from a zero resolution to a non-zero resolution in response to a determination that a maximum possible resolution in a previous level is less than a non-zero resolution of a desired state.
- the method may also include monitoring transitions from a non-zero resolution to a zero resolution for each path in the trellis; setting a variable to a predefined value when a transition from the non-zero resolution to the zero resolution occurs; and selecting a current branch as a branch with lowest cost that enters a current state.
- the method may further include defining combinations of directions for multidirectional cameras as states on each branch of the trellis and performing an exhaustive search by: examining all eligible combinations, computing a cost of each combination, and selecting a combination with a lowest cost.
- the method may include evaluating the points and possible combinations of camera assignment to the points; upon determining one or more possible camera and resolution combinations for each point, examining a resolution consistency; omitting possible camera and resolution combinations that violate the resolution consistency; and selecting a camera and resolution combination among remaining camera and resolution combinations.
- the method may also include evaluating the cameras and testing possible resolutions for each camera without examining a resolution consistency; setting a resolution for a camera to zero in response to a determination that the camera is to be turned off of set to e resolution that fails to cover a point to which the camera is assigned; and/or defining a total cost of each transition as a linear combination of a number of cameras and resolution costs.
- the cameras may include one of: a single resolution camera, a multi-resolution camera, a multidirectional camera, a multi-resolution/multidirectional camera, and a pan-tilt-zoom (PTZ) camera.
- a computing device for optimal camera selection in array of cameras for monitoring and surveillance applications.
- the computing device may include a memory configured to store instructions and a processor.
- the processor may be configured to determine a plurality of resolutions associated with a plurality of cameras defined for intervals along a linear axis; receive information associated with points on the intervals and desired resolutions for the points; form a combinatorial state trellis, where each level represents a point according to a linear order of the points and possible combinations of camera resolutions covering the point are listed as states on a corresponding level; and evaluate optimal paths through the levels while obeying resolution constraints in each path that is traversed in the trellis until a survival path is determined.
- the processor may be further configured to form transitions from one level to a next level, while following a resolution consistency for common cameras corresponding to a start and an end state of each transition; enable a transition from a non-zero resolution to a zero resolution in response to a determination that a maximum possible resolution in the next level is less than a resolution at a start of the transition; and/or enable a transition from a zero resolution to a non-zero resolution in response to a determination that a maximum possible resolution in a previous level is less than a non-zero resolution of a desired state.
- the processor may also be configured to monitor transitions from a non-zero resolution to a zero resolution for each path in the trellis; set a variable to a predefined value when a transition from the non-zero resolution to the zero resolution occurs; and select a current branch as a branch with lowest cost that enters a current state.
- the processor may further define combinations of directions for multidirectional cameras as states on each branch of the trellis and perform an exhaustive search by examining all eligible combinations, computing a cost of each combination, and selecting a combination with a lowest cost.
- the processor may evaluate the points and possible combinations of camera assignment to the points; upon determining one or more possible camera and resolution combinations for each point, examine a resolution consistency; omit possible camera and resolution combinations that violate the resolution consistency; and select a camera and resolution combination among remaining camera and resolution combinations.
- the processor may also evaluate the cameras and testing possible resolutions for each camera without examining a resolution consistency and/or set a resolution for a camera to zero in response to a determination that the camera is to be turned off of set to e resolution that fails to cover a point to which the camera is assigned.
- the processor may further define a total cost of each transition as a linear combination of a number of cameras and resolution costs.
- the cameras may include one of: a single resolution camera, a multi-resolution camera, a multidirectional camera, a multi-resolution/multidirectional camera, and a pan-tilt-zoom (PTZ) camera.
- a method for optimal light subset selection in a lighting array that achieves a desired intensity for an area of illumination.
- An example method may include determining a plurality of lighting intensities associated with a plurality of lights defined for intervals along a linear axis; receiving information associated with points on the intervals and desired lighting intensities for the points; forming a combinatorial state trellis, wherein each level represents a point according to a linear order of the points and possible combinations of lighting intensities covering the point are listed as states on a corresponding level; and evaluating optimal paths through the levels while obeying lighting intensity constraints in each path that is traversed in the trellis until a survival path is determined.
- a computing device for optimal light subset selection in a lighting array that achieves a desired intensity for an area of illumination.
- the computing device may include a memory configured to store instructions and a processor.
- the processor may be configured to determine a plurality of lighting intensities associated with a plurality of lights defined for intervals along a linear axis; receive information associated with points on the intervals and desired lighting intensities for the points; form a combinatorial state trellis, wherein each level represents a point according to a linear order of the points and possible combinations of lighting intensities covering the point are listed as states on a corresponding level; and evaluate optimal paths through the levels while obeying lighting intensity constraints in each path that is traversed in the trellis until a survival path is determined.
- a computer readable storage medium with instructions stored thereon for executing the above methods at one or more processors for optimizing an efficiency of camera placement, numbers, and resolution in a multi-camera monitoring environment may also be described.
- the implementer may opt for a mainly hardware and/or firmware vehicle; if flexibility is paramount, the implementer may opt for a mainly software implementation; or, yet again alternatively, the implementer may opt for some combination of hardware, software, and/or firmware.
- Examples of a signal bearing medium include, but are not limited to, the following: a recordable type medium such as floppy disk, a hard disk drive, a Compact Disc (CD), a Digital Versatile Disk (DVD), a digital tape, a computer memory, a solid state drive, etc.; and a transmission type medium such as a digital and/or an analog communication medium (e.g., a fiber optic cable, a waveguide, a wired communications link, a wireless communication link, etc.).
- a recordable type medium such as floppy disk, a hard disk drive, a Compact Disc (CD), a Digital Versatile Disk (DVD), a digital tape, a computer memory, a solid state drive, etc.
- a transmission type medium such as a digital and/or an analog communication medium (e.g., a fiber optic cable, a waveguide, a wired communications link, a wireless communication link, etc.).
- a typical data processing system generally includes one or more of a system unit housing, a video display device, a memory such as volatile and non-volatile memory, processors such as microprocessors and digital signal processors, computational entities such as operating systems, drivers, graphical user interfaces, and applications programs, one or more interaction devices, such as a touch pad or screen, and/or control systems including feedback loops and control motors (e.g., feedback for sensing position and/or velocity of gantry systems; control motors for moving and/or adjusting components and/or quantities).
- a typical data processing system may be implemented utilizing any suitable commercially available components, such as those typically found in data computing/communication and/or network computing/communication systems.
- the herein described subject matter sometimes illustrates different components contained within, or connected with, different other components. It is to be understood that such depicted architectures are merely exemplary, and that in fact many other architectures may be implemented which achieve the same functionality. In a conceptual sense, any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved. Hence, any two components herein combined to achieve a particular functionality may be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermediate components.
- any two components so associated may also be viewed as being “operably connected”, or “operably coupled”, to each other to achieve the desired functionality, and any two components capable of being so associated may also be viewed as being “operably couplable”, to each other to achieve the desired functionality.
- operably couplable include but are not limited to physically connectable and/or physically interacting components and/or wirelessly interactable and/or wirelessly interacting components and/or logically interacting and/or logically interactable components.
- a range includes each individual member.
- a group having 1-3 cells refers to groups having 1, 2, or 3 cells.
- a group having 1-5 cells refers to groups having 1, 2, 3, 4, or 5 cells, and so forth.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
- Closed-Circuit Television Systems (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Camera Bodies And Camera Details Or Accessories (AREA)
Abstract
Description
X=[X 1 X 2 . . . X NoC] [1]
Assuming a maximum resolution value is nR, each vector element may be considered as
Xiε{0, 1, 2, . . . , n R} [2]
in a general case, where points and cameras may be distributed over a two-dimensional area at any location.
Res assj(X)=max(X cj), [3]
where
C j ={i,X i ≦V(i,j)}. [4]
Alternatively, following equation may be written using vector inequalities:
Res assj(X)=max(X(X≦V(i,j))). [5]
The optimization is N-P complete. Thus, a greedy technique may be employed for computing the optimization in some examples.
Resolutions cost=L*Res desint−Σi=1 len i *Res assi=Σi=1 len i*(Res desint −Res assi), [10]
where L and Resdesint are the length and desired resolution of the original interval respectively. Also, leni and Resassi are the length of and the assigned resolution to each interval respectively. Minimizing the above cost may be substantially equal to minimizing the term in the parentheses for every value of i. Since the cost is a linear combination of sub-costs with positive coefficients, in order to minimize the overall cost, the cost of each sub-interval may be independently minimized.
Claims (12)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/115,093 US9591272B2 (en) | 2012-04-02 | 2013-03-28 | Optimal camera selection in array of monitoring cameras |
US15/422,456 US9942468B2 (en) | 2012-04-02 | 2017-02-01 | Optimal camera selection in array of monitoring cameras |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261618925P | 2012-04-02 | 2012-04-02 | |
PCT/CA2013/050260 WO2013149340A1 (en) | 2012-04-02 | 2013-03-28 | Optimal camera selection iν array of monitoring cameras |
US14/115,093 US9591272B2 (en) | 2012-04-02 | 2013-03-28 | Optimal camera selection in array of monitoring cameras |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CA2013/050260 A-371-Of-International WO2013149340A1 (en) | 2012-04-02 | 2013-03-28 | Optimal camera selection iν array of monitoring cameras |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/422,456 Division US9942468B2 (en) | 2012-04-02 | 2017-02-01 | Optimal camera selection in array of monitoring cameras |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140055621A1 US20140055621A1 (en) | 2014-02-27 |
US9591272B2 true US9591272B2 (en) | 2017-03-07 |
Family
ID=49299895
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/115,093 Expired - Fee Related US9591272B2 (en) | 2012-04-02 | 2013-03-28 | Optimal camera selection in array of monitoring cameras |
US15/422,456 Expired - Fee Related US9942468B2 (en) | 2012-04-02 | 2017-02-01 | Optimal camera selection in array of monitoring cameras |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/422,456 Expired - Fee Related US9942468B2 (en) | 2012-04-02 | 2017-02-01 | Optimal camera selection in array of monitoring cameras |
Country Status (3)
Country | Link |
---|---|
US (2) | US9591272B2 (en) |
JP (1) | JP5992090B2 (en) |
WO (1) | WO2013149340A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11108994B2 (en) * | 2018-08-27 | 2021-08-31 | Hitachi Kokusai Electric Inc. | Image display system and image display method |
Families Citing this family (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6349637B2 (en) * | 2013-07-11 | 2018-07-04 | 株式会社デンソー | Image synthesizer for vehicles |
US9426365B2 (en) * | 2013-11-01 | 2016-08-23 | The Lightco Inc. | Image stabilization related methods and apparatus |
US9729857B2 (en) * | 2014-04-08 | 2017-08-08 | Semyon Nisenzon | High resolution depth map computation using multiresolution camera clusters for 3D image generation |
US9984544B2 (en) * | 2015-02-17 | 2018-05-29 | Sap Se | Device layout optimization for surveillance devices |
GB2535706A (en) | 2015-02-24 | 2016-08-31 | Nokia Technologies Oy | Device with an adaptive camera array |
US10230326B2 (en) | 2015-03-24 | 2019-03-12 | Carrier Corporation | System and method for energy harvesting system planning and performance |
EP3274932A1 (en) | 2015-03-24 | 2018-01-31 | Carrier Corporation | Integrated system for sales, installation, and maintenance of building systems |
US10928785B2 (en) | 2015-03-24 | 2021-02-23 | Carrier Corporation | Floor plan coverage based auto pairing and parameter setting |
US11036897B2 (en) | 2015-03-24 | 2021-06-15 | Carrier Corporation | Floor plan based planning of building systems |
EP3275204B1 (en) | 2015-03-24 | 2020-07-22 | Carrier Corporation | System and method for capturing and analyzing multidimensional building information |
CN107660300B (en) | 2015-03-24 | 2021-01-29 | 开利公司 | System and method for providing a graphical user interface indicating intruder threat levels for a building |
US10944837B2 (en) | 2015-03-24 | 2021-03-09 | Carrier Corporation | Floor-plan based learning and registration of distributed devices |
US10756830B2 (en) | 2015-03-24 | 2020-08-25 | Carrier Corporation | System and method for determining RF sensor performance relative to a floor plan |
US10326981B2 (en) * | 2015-05-15 | 2019-06-18 | Semyon Nisenzon | Generating 3D images using multi-resolution camera set |
US11188034B2 (en) * | 2015-07-31 | 2021-11-30 | Dallmeier Electronic Gmbh & Co. Kg | System for monitoring and influencing objects of interest and processes carried out by the objects, and corresponding method |
WO2017038450A1 (en) * | 2015-09-02 | 2017-03-09 | 日本電気株式会社 | Monitoring system, monitoring network construction method, and program |
US10469821B2 (en) * | 2016-06-17 | 2019-11-05 | Altek Semiconductor Corp. | Stereo image generating method and electronic apparatus utilizing the method |
CN108471496B (en) * | 2018-02-13 | 2020-11-03 | 深圳市瑞立视多媒体科技有限公司 | Camera configuration method and device |
US10771715B2 (en) | 2016-08-09 | 2020-09-08 | Shenzhen Realis Multimedia Technology Co., Ltd. | Camera configuration method and apparatus |
CN108495057B (en) * | 2018-02-13 | 2020-12-08 | 深圳市瑞立视多媒体科技有限公司 | Camera configuration method and device |
CN108449551B (en) * | 2018-02-13 | 2020-11-03 | 深圳市瑞立视多媒体科技有限公司 | Camera configuration method and device |
EP3490245A4 (en) * | 2016-08-09 | 2020-03-11 | Shenzhen Realis Multimedia Technology Co., Ltd. | Camera configuration method and device |
DE102016223132A1 (en) * | 2016-11-23 | 2018-05-24 | Robert Bosch Gmbh | Method and system for detecting a raised object located within a parking lot |
US10089751B1 (en) * | 2017-05-31 | 2018-10-02 | Booz Allen Hamilton Inc. | Video processing technique for 3D target location identification |
US11738785B2 (en) * | 2018-02-15 | 2023-08-29 | Yaakov Frucht | System and method for detecting an intruder on tracks |
US20190248390A1 (en) * | 2018-02-15 | 2019-08-15 | Hi-Tec Security Systems Ltd. | Track intrusion detection system |
EP3672231B1 (en) * | 2018-12-21 | 2021-05-26 | Axis AB | Adaptive storage between multiple cameras in a video recording system |
US10957074B2 (en) | 2019-01-29 | 2021-03-23 | Microsoft Technology Licensing, Llc | Calibrating cameras using human skeleton |
FR3105865B1 (en) * | 2019-12-31 | 2022-03-11 | Data Smart Process | Method and system for deploying surveillance cameras |
FR3105852A1 (en) * | 2019-12-31 | 2021-07-02 | Data Smart Process | Surveillance camera deployment method and system |
JP2022099017A (en) * | 2020-12-22 | 2022-07-04 | 日本電気株式会社 | Test support device, test support method and program |
CN114900602B (en) * | 2022-06-08 | 2023-10-17 | 北京爱笔科技有限公司 | Method and device for determining video source camera |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2003030550A1 (en) | 2001-09-27 | 2003-04-10 | Koninklijke Philips Electronics N.V. | Optimal multi-camera setup for computer-based visual surveillance |
US6741250B1 (en) | 2001-02-09 | 2004-05-25 | Be Here Corporation | Method and system for generation of multiple viewpoints into a scene viewed by motionless cameras and for presentation of a view path |
US20050065802A1 (en) | 2003-09-19 | 2005-03-24 | Microsoft Corporation | System and method for devising a human interactive proof that determines whether a remote client is a human or a computer program |
US6954224B1 (en) | 1999-04-16 | 2005-10-11 | Matsushita Electric Industrial Co., Ltd. | Camera control apparatus and method |
US20050225638A1 (en) | 1997-01-28 | 2005-10-13 | Canon Kabushiki Kaisha | Apparatus and method for controlling a camera based on a displayed image |
US20060222209A1 (en) | 2005-04-05 | 2006-10-05 | Objectvideo, Inc. | Wide-area site-based video surveillance system |
WO2006132029A1 (en) | 2005-06-07 | 2006-12-14 | Matsushita Electric Industrial Co., Ltd. | Monitoring system, monitoring method, and camera terminal |
US20070039030A1 (en) | 2005-08-11 | 2007-02-15 | Romanowich John F | Methods and apparatus for a wide area coordinated surveillance system |
US7187402B2 (en) | 1995-12-19 | 2007-03-06 | Canon Kabushiki Kaisha | Communication apparatus, image processing apparatus, communication method, and image processing method |
US7440594B2 (en) | 2002-07-30 | 2008-10-21 | Omron Corporation | Face identification device and face identification method |
US20090040367A1 (en) | 2002-05-20 | 2009-02-12 | Radoslaw Romuald Zakrzewski | Method for detection and recognition of fog presence within an aircraft compartment using video images |
US7668345B2 (en) | 2005-03-31 | 2010-02-23 | Hitachi, Ltd. | Image processing apparatus, image processing system and recording medium for programs therefor |
JP2010054203A (en) | 2008-08-26 | 2010-03-11 | Panasonic Electric Works Co Ltd | Luminaire |
EP2328131A2 (en) | 2005-03-25 | 2011-06-01 | Sensormatic Electronics LLC | Intelligent camera selection and object tracking |
US20120038776A1 (en) | 2004-07-19 | 2012-02-16 | Grandeye, Ltd. | Automatically Expanding the Zoom Capability of a Wide-Angle Video Camera |
US20120307067A1 (en) * | 2011-06-01 | 2012-12-06 | Honeywell International Inc. | System and method for automatic camera placement |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6359647B1 (en) * | 1998-08-07 | 2002-03-19 | Philips Electronics North America Corporation | Automated camera handoff system for figure tracking in a multiple camera system |
US8432449B2 (en) * | 2007-08-13 | 2013-04-30 | Fuji Xerox Co., Ltd. | Hidden markov model for camera handoff |
-
2013
- 2013-03-28 WO PCT/CA2013/050260 patent/WO2013149340A1/en active Application Filing
- 2013-03-28 US US14/115,093 patent/US9591272B2/en not_active Expired - Fee Related
- 2013-03-28 JP JP2015502029A patent/JP5992090B2/en not_active Expired - Fee Related
-
2017
- 2017-02-01 US US15/422,456 patent/US9942468B2/en not_active Expired - Fee Related
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7187402B2 (en) | 1995-12-19 | 2007-03-06 | Canon Kabushiki Kaisha | Communication apparatus, image processing apparatus, communication method, and image processing method |
US20050225638A1 (en) | 1997-01-28 | 2005-10-13 | Canon Kabushiki Kaisha | Apparatus and method for controlling a camera based on a displayed image |
US6954224B1 (en) | 1999-04-16 | 2005-10-11 | Matsushita Electric Industrial Co., Ltd. | Camera control apparatus and method |
US6741250B1 (en) | 2001-02-09 | 2004-05-25 | Be Here Corporation | Method and system for generation of multiple viewpoints into a scene viewed by motionless cameras and for presentation of a view path |
WO2003030550A1 (en) | 2001-09-27 | 2003-04-10 | Koninklijke Philips Electronics N.V. | Optimal multi-camera setup for computer-based visual surveillance |
US20090040367A1 (en) | 2002-05-20 | 2009-02-12 | Radoslaw Romuald Zakrzewski | Method for detection and recognition of fog presence within an aircraft compartment using video images |
US7440594B2 (en) | 2002-07-30 | 2008-10-21 | Omron Corporation | Face identification device and face identification method |
US20050065802A1 (en) | 2003-09-19 | 2005-03-24 | Microsoft Corporation | System and method for devising a human interactive proof that determines whether a remote client is a human or a computer program |
US20120038776A1 (en) | 2004-07-19 | 2012-02-16 | Grandeye, Ltd. | Automatically Expanding the Zoom Capability of a Wide-Angle Video Camera |
EP2328131A2 (en) | 2005-03-25 | 2011-06-01 | Sensormatic Electronics LLC | Intelligent camera selection and object tracking |
US7668345B2 (en) | 2005-03-31 | 2010-02-23 | Hitachi, Ltd. | Image processing apparatus, image processing system and recording medium for programs therefor |
US20060222209A1 (en) | 2005-04-05 | 2006-10-05 | Objectvideo, Inc. | Wide-area site-based video surveillance system |
WO2006132029A1 (en) | 2005-06-07 | 2006-12-14 | Matsushita Electric Industrial Co., Ltd. | Monitoring system, monitoring method, and camera terminal |
US20070039030A1 (en) | 2005-08-11 | 2007-02-15 | Romanowich John F | Methods and apparatus for a wide area coordinated surveillance system |
JP2010054203A (en) | 2008-08-26 | 2010-03-11 | Panasonic Electric Works Co Ltd | Luminaire |
US20120307067A1 (en) * | 2011-06-01 | 2012-12-06 | Honeywell International Inc. | System and method for automatic camera placement |
Non-Patent Citations (40)
Title |
---|
"Object Tracking & Understanding," Mitsubishi Electric Research Laboratories, accessed at http://web.archive.org/web/20081120032302/http://www.merl.com/projects/ObjectTracking/, Modified on Jul. 15, 2004, p. 1-1. |
Aghdasi et al., "High-Resolution Images with Minimum Energy Dissipation and Maximum Field-of-view in Camera-Based Wireless Multimedia Sensor Networks", Sensors, Aug. 19, 2009. |
Akyildiz et al., "A survey on wireless multimedia sensor networks", Computer Networks, vol. 51, No. 4, pp. 921-960, Published on Oct. 5, 2007. |
Arampatzis et al., "A survey of applications of wireless sensors and wireless sensor networks," Intelligent Control, 2005. Proceedings of the 2005 IEEE International Symposium on, Mediterranean Conference on Control and Automation, Date of Conference: Jun. 27-29, 2005. |
Arulampalam. M.S., et al., "A Tutorial on Particle Filters for Online Nonlinear/Nun-Gaussian Bayesian Tracking," IEEE Transactions on Signal Processing, vol. 50, Issue 2, pp. 174-188 (Feb. 2002). |
Cai et al., "Target-oriented scheduling in directional sensor networks," in INFOCOM 2007. 26th IEEE International Conference, Date of Conference: May 6-12, 2007. |
Charfi et al., "Challenging issues in visual sensor networks," Wireless Communications, IEEE, vol. 16, No. 2, pp. 44-49, Published on Apr. 2009. |
Cheng et al., "Distributed barrier coverage in wireless visual sensor networks with B-QoM," Sensors Journal, IEEE, vol. 12, No. 6, pp. 1726-1735, Jun. 2011. |
Comaniciu, D., et al "Kernel-Based Object Tracking" IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, Issue 5, pp. 564-577 (May 2003). |
Dieber et al., "Resource-aware coverage and task assignment in visual sensor networks," Circuits and Systems for video Technology, IEEE Transactions on, vol. 21, No. 10, pp. 1424-1437, Published on Oct. 2011. |
Fusco et al., "Selection and orientation of directional sensors for coverage maximization," in Sensor Mesh and Ad Hoc Communications and Networks, 2009. SECON ″09. 6th Annual IEEE Communications Society Conference on Jun. 2009. |
Fusco et al., "Selection and orientation of directional sensors for coverage maximization," in Sensor Mesh and Ad Hoc Communications and Networks, 2009. SECON ''09. 6th Annual IEEE Communications Society Conference on Jun. 2009. |
Guvensan et a.l, "On coverage issues in directional sensor networks: A survey," Ad Hoc Networks, vol. 9, No. 7, pp. 1238-1255, Published on Sep. 2011. |
Huang et al., "The coverage problem in a wireless sensor network," in Proceedings of the 2nd ACM international conference on Wireless sensor networks and applications, ser. WSNA '03. New York, NY, Published on Sep. 19, 2003. |
International Search Report and Written Opinion for PCT/CA2013/050260 filed Mar. 28, 2013, mailed Jul. 22, 2013. |
J. Zhao and S.S. Cheung, "Multi-Camera Surveillance with Visual Tagging and Generic Camera Placement", 1st ACM/IEEE International Conference on Distributed Smart Cameras, 2007. * |
J. Zhao and S.S. Cheung, "Multi-Camera Surveillance with Visual Tagging and Generic Camera Placement", First ACM/IEEE International Conference on Distributed Smart Cameras, 2007. * |
Kandoth et al., "Angular mobility assisted coverage in directional sensor networks," in Network Based Information Systems, 2009. NBIS 2009, Aug. 2009. |
Kim et al., "Image Quality and Lifetime Co-optimization in Wireless Multi-Camera Systems", Circuits and Systems (ISCAS), pp. 2641,2644, May 2011. |
Krahnstoever, N., et al., "Collaborative Real-Time Control of Active Cameras in Large Scale Surveillance Systems," M2SFA2 2008: Workshop on Multi-camera and Multi-modal Sensor Fusion, pp. 12 (Oct. 13, 2008). |
Lobaton et al "A distributed topological camera network representation for tracking applications," Image Processing, IEEE Transactions on, vol. 19, No. 10, pp. 2516-2529, Oct. 2010. |
Lobaton et al., "Algebraic approach to recovering topological information in distributed camera networks," in Information Processing in Sensor Networks, 2009. IPSN 2009, Apr. 2009. |
Ma et al., "A coverage-enhancing method for 3D directional sensor networks," in INFOCOM 2009, IEEE,, pp. 2791-2795, Published on Apr. 2009. |
Mavrinac et al., "Task-Oriented Optimal View Selection in a Calibrated Multi-Camera System", Advanced Intelligent Mechatronics (AIM), pp. 69-74, Jul. 2012. |
Micheloni, C., et al., "A network of co-operative cameras for visual surveillance," IEE Proceedings Vision, Image and Signal Processing, , vol. 152, Issue 2, pp. 205-212 (Apr. 8, 2005). |
Park et al., "A look-up table based approach for solving the camera selection problem in large camera networks," in in Proceedings of the International Workshop on Distributed Smart Cameras, Oct. 2006. |
Rekleitis et al., "Simultaneous planning localization, and mapping in a camera sensor network," Robotics and Autonomous Systems (RAS) Journal, special, vol. 54, pp. 921-932, 2006, Aug. 2006. |
S. Indu, S. Chaudhury, N.R. Mittal and A. Bhattacharyya, "Optimal Sensor Placement for Surveillance of Large Spaces", Third International Conference on Digital Object Identifier, 2009. * |
Shen et al., "A multi-camera surveillance system that estimates quality-of-view measurement," in Image Processing, 2007. ICIP 2007. IEEE International Conference, Oct. 2007. |
Soro et al., "A Survey of Visual Sensor Networks", Advances in Multimedia, Accepted May 13, 2009. |
Soro et al., "Camera selection in visual sensor networks", In Advanced video and Signal Based Surveillance, 2007. AVSS 2007. IEEE, Sep. 2007. |
Tseng et al., "k-Angle Object Coverage Problem in a Wireless Sensor Network," Sensors Journal, IEEE, vol. PP, No. 99, p. 1, Dec. 2012. |
Viola, and Jones, M., "Rapid Object Detection Using a Boosted Cascade of Simple Features," Mitsubishi Electric Research Laboratories, pp. 13 (May 2004). |
Viterbi et al., "Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding Algorithm," Information Theory, IEEE Transactions on, 13, No. 2, pp. 260-269, Apr. 1967. |
Wang et al., "Priority-based target coverage in directional sensor networks using a genetic algorithm," Computers and Mathematics with Applications, vol. 57,No. 11-12, pp. 1915-1922, Jun. 2009. |
Wolf et al., "Finding the Best Set of K Paths Through a Trellis With Application to Multitarget Tracking", IEEE Transactions on Aerospace and Electronic Systems, vol. 25, Issue 2, pp. 287-295, ISSN#OO I 8-9251, Digital Object Identifier I 0-1 I 0917.18692; (Wolf et al.), published on Mar. 19, 1989. |
Wu, et al., "Efficient algorithms for probabilistic k-coverage in directional sensor networks," in Intelligent Sensors, Sensor Networks and Information Processing, 2008. ISSNIP 2008, Dec. 2008. |
Yang et al., "Coverage quality based target-oriented scheduling in directional sensor networks," in Communications (ICC), 2010 IEEE International Conference on May 2010. |
Yifan Zhou, Jenny Benois-Pineau, Henri Nicolas, "A multi-resolution particle filter tracking with a dual consistency check for model update in a multi-camera environment", 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS, 2010. * |
Youn-Hee Han, Chan-Myung Kim and Joon-Min Gil, "A Greedy Algorithm for Target Coverage Scheduling in Directional Sensor Networks", 5th International Conference on Ubiquitous Information Technologies and Applications (CUTE), 2010. * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11108994B2 (en) * | 2018-08-27 | 2021-08-31 | Hitachi Kokusai Electric Inc. | Image display system and image display method |
Also Published As
Publication number | Publication date |
---|---|
US9942468B2 (en) | 2018-04-10 |
JP2015517247A (en) | 2015-06-18 |
US20170150036A1 (en) | 2017-05-25 |
US20140055621A1 (en) | 2014-02-27 |
WO2013149340A1 (en) | 2013-10-10 |
JP5992090B2 (en) | 2016-09-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9942468B2 (en) | Optimal camera selection in array of monitoring cameras | |
US10664310B2 (en) | Memory access optimisation using per-layer computational mapping and memory allocation for CNN application | |
JP7328379B2 (en) | Method, system, and apparatus for improving convolution efficiency | |
US20220156628A1 (en) | Software-defined quantum computer | |
EP3501010B1 (en) | Rendering operations using sparse volumetric data | |
EP3612990B1 (en) | Power-efficient deep neural network module configured for layer and operation fencing and dependency management | |
US20230165643A1 (en) | Surgical workflow and activity detection based on surgical videos | |
US11954583B2 (en) | Transposed convolution using systolic array | |
US20190362235A1 (en) | Hybrid neural network pruning | |
US20220292163A1 (en) | Dilated convolution using systolic array | |
US9424079B2 (en) | Iteration support in a heterogeneous dataflow engine | |
US9498715B2 (en) | Speculative rendering using historical player data | |
US10684776B2 (en) | Memory configuration for inter-processor communication in an MPSoC | |
KR20180050928A (en) | Convolutional neural network processing method and apparatus | |
KR20180073118A (en) | Convolutional neural network processing method and apparatus | |
US12073199B2 (en) | Reducing computation in neural networks using self-modifying code | |
US11501145B1 (en) | Memory operation for systolic array | |
CN110287855A (en) | A kind of pedestrian track acquisition methods and system | |
Velez et al. | A reconfigurable embedded vision system for advanced driver assistance | |
US20210304010A1 (en) | Neural network training under memory restraint | |
CN114761920A (en) | Hardware accelerator with reconfigurable instruction set | |
US11403206B2 (en) | Method and apparatus for debugging, and system on chip | |
Gurcan et al. | Heterogeneous CPU–GPU tracking–learning–detection (H-TLD) for real-time object tracking | |
CN107832228A (en) | A kind of test case reduction method, device, equipment and storage medium | |
Zhuang et al. | Towards high-quality CGRA mapping with graph neural networks and reinforcement learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MCMASTER UNIVERSITY, CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHIRANI, SHAHRAM;SAMAVI, SHADROKH;SOROUSHMEHR, SAYEDMOHAMMADREZA;AND OTHERS;SIGNING DATES FROM 20130426 TO 20130508;REEL/FRAME:031525/0179 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
AS | Assignment |
Owner name: CRESTLINE DIRECT FINANCE, L.P., TEXAS Free format text: SECURITY INTEREST;ASSIGNOR:EMPIRE TECHNOLOGY DEVELOPMENT LLC;REEL/FRAME:048373/0217 Effective date: 20181228 |
|
AS | Assignment |
Owner name: EMPIRE TECHNOLOGY DEVELOPMENT LLC, WASHINGTON Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CRESTLINE DIRECT FINANCE, L.P.;REEL/FRAME:049924/0794 Effective date: 20190501 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20210307 |