3D Height / Width Sample Clauses

3D Height / Width. Building upon the height and width measurements taken in the image plane, height and width measurements are now taken incorporating known camera geometry. This method allows known camera positions relative to object positions in the scene to compute very accurate object height and width measurements. These measurements are invariant to perspective or affine distortion, but not to occlusion. It is important to note, that although the measurements themselves are invariant to 3D rotation within the scene, providing accurate measurements of the object irrespective of their position, this rotation will alter the size of the object under observation. For example, a car viewed side on will be a different width to a car viewed head on. This change can be compensated for within the model created, as the centroid positions will consider these variations. However, this may reduce the accuracy of the final classification, as some classes may overlap. This overlap may mar the decision boundaries. Region moment representations interpret a normalised gray­level image function as a probability density of a 2D random variable. Properties of this random variable can be described using statistical characteristics ­ moments [21]. Moments allow an object to be described using standard statistics, for example, the objects area ( m00 ). The use of several statistical descriptions of an object allows a specific object's type (classification) to be postulated. This system has been used to great effect in object and letter classification by ▇▇▇▇▇▇▇ and Suk [23]. In digitised images, moments are created such that: mpq = ∑ i p jq f  i , j  (5.1) Where i, j are the pixel co­ordinates and p,q describe the moments to create (for example, to create m00 , p and q = 0). It is advantageous for moments to be as invariant as possible as in Section 5.1.2.1., where the heights to width measurements are made invariant to perspective projection. Invariance in moments allows for a more robust and flexible object representation. Translation invariance (i.e. moments that appear the same at any position within the image) can be achieved if image central moments are used [22], through the generalised formula: μ pq = ∑
3D Height / Width. Encouraged by the non­invariant model of height and width (size), an invariant form was created using known camera geometry to increase measurement accuracy and overcome the problems of distance, or perspective, within the data files. This model used the same decision­boundary creation schemes as the non­invariant model, and matched in the same manner, but scores over the non­invariant model in the accuracy of the measurements returned. Although this invariant model gives better results than its non­invariant form (as can be seen from the results obtained), the classification accuracy is not high enough for outright object class to be determined based on this technique. This is mainly due to the problems of 3D rotation within the invariant model, for although it is invariant to scale, the measurements returned are non­linear (i.e. as an object turns, its class remains the same, but its size alters non­linearly). This type of size­based model cannot easily cope with this problem. The model used can claim to reduce the effects of 3D rotation, which is done using Mahalanobis distance, which takes the distribution (alterations) in rotation into account for each object. Unfortunately, this is obviously not sufficient for the task, however, based on the results listed in [85].