sensors on 3d digitization full report
#1


[attachment=1279][attachment=1280]


ABSTRACT

Machine vision involves the analysis of the properties of the luminous flux reflected or radiated by objects. To recover the geometrical structures of these objects, either to recognize or to measure their dimension, two basic vision strategies are available.
The first strategy, known as passive vision, attempts to analyze the structure of the scene under ambient light. In contrast, the second, known as active vision, attempts to reduce the way in which images are formed. Sensors that capitalize on active vision can resolve most of the ambiguities found with 2-D imaging systems. Moreover, with laser “based approaches, the 3-D information becomes relatively insensitive to background illumination and surface texture. Complete images of visible surfaces that are rather featureless to the human eye or a video camera can be generated. Thus the task of processing 3-D data is greatly simplified.

INTRODUCTION
Digital 3D imaging can benefit from advances in VLSI technology in order to accelerate its deployment in many fields like visual communication and industrial automation. High-resolution 3D images can be acquired using laser-based vision systems. With this approach, the 3D information becomes relatively insensitive to background illumination and surface texture. Complete images of visible surfaces that are rather featureless to the human eye or a video camera can be generated. Intelligent digitizers will be capable of measuring accurately and simultaneously colour and 3D.

COLOUR 3D IMAGING TECHNOLOGY

Machine vision involves the analysis of the properties of the luminous flux reflected or radiated by objects. To recover the geometrical structures of these objects, either to recognize or to measure their dimension, two basic vision strategies are available [1].
Passive vision, attempts to analyze the structure of the scene under ambient light. [1] Stereoscopic vision is a passive optical technique. The basic idea is that two or more digital images are taken from known locations. The images are then processed to find the correlations between them. As soon as matching points are identified, the geometry can be computed.
Active vision attempts to reduce the ambiguity of scene analysis by structuring the way in which images are formed. Sensors that capitalize on active vision can resolve most of the ambiguities found with two-dimensional imaging systems. Lidar based or triangulation based laser range cameras are examples of active vision technique. One digital 3D imaging system based on optical triangulation were developed and demonstrated.

AUTOSYNCHRONIZED SCANNER

The auto-synchronized scanner, depicted schematically on Figure 1, can provide registered range and colour data of visible surfaces. A 3D surface map is captured by scanning a laser spot onto a scene, collecting the reflected laser light, and finally focusing the beam onto a linear laser spot sensor. Geometric and photometric corrections of the raw data give two images in perfect registration: one with x, y, z co-ordinates and a second with reflectance data. The laser beam composed of multiple visible wavelengths is used for the purpose of measuring the colour map of a scene (reflectance map).
Advantage
Triangulation is the most precise method of 3D
Limitation
Increasing the accuracy increases the triangulation distance. The larger the triangulation distance, the more shadows appear on the scanning object and the scanning head must be made larger.
SENSORS FOR 3D IMAGING
The sensors used in the autosynchronized scanner include
1. SYNCHRONIZATION CIRCUIT BASED UPON DUAL PHOTOCELLS
This sensor ensures the stability and the repeatability of range measurements in environment with varying temperature. Discrete implementations of the so-called synchronization circuits have posed many problems in the past. A monolithic version of an improved circuit has been built to alleviate those problems. [1]
2. LASER SPOT POSITION MEASUREMENT SENSORS
High-resolution 3D images can be acquired using laser-based vision systems. With this approach, the 3D information becomes relatively insensitive to background illumination and surface texture. Complete images of visible surfaces that are rather featureless to the human eye or a video camera can be generated.
POSITION SENSITIVE DETECTORS
Many devices have been built or considered in the past for measuring the position of a laser spot more efficiently. One method of position detection uses a video camera to capture an image of an object electronically. Image processing techniques are then used to determine the location of the object .For situations requiring the location of a light source on a plane, a position sensitive detector (PSD) offers the potential for better resolution at a lower system cost.

The PSD is a precision semiconductor optical sensor which produces output currents related to the centre of mass of light incident on the surface of the device.

While several design variations exist for PSDs, the basic device can be described as a large area silicon p-i-n photodetector. The detectors can be available in single axis and dual axis models.

DUAL AXIS PSD

This particular PSD is a five terminal device bounded by four collection surfaces; one terminal is connected to each collection surface and one provides a common return. Photocurrent is generated by light which falls on the active area of the PSD will be collected by these four perimeter electrodes.

The amount of current flowing between each perimeter terminal and the common return is related to the proximity of the centre of mass of the incident light spot to each collection surface. The difference between the up current and the down current is proportional to the Y-axis position of the light spot .Similarly ,the right current minus the left current gives the X-axis position. The designations up,down, rightand Ëœleftare arbitrary; the device may be operated in any relative spatial orientation.
LASER SENSORS FOR 3D IMAGING
The state of the art in laser spot position sensing methods can be divided into two broad classes according to the way the spot position is sensed. Among those, one finds continuous response position sensitive detectors (CRPSD) and discrete response position sensitive detectors (DRPSD)
CONTINOUS RESPONSE POSITION SENSITIVE DETECTORS (CRPSD)
The category CRPSD includes lateral effect photodiode .

In basic structure of a p-n type single axis lateral effect photodiode. Carriers are produced by light impinging on the device are separated in the depletion region and distributed to the two sensing electrodes according to the Ohmâ„¢s law. Assuming equal impedances, The electrode that is the farthest from the centroid of the light distribution collects the least current. The normalized position of the centroid when the light intensity fluctuates is given by P =I2-I1/I1+I2.
The actual position on the detector is found by multiplying P by d/2 ,where d is the distance between the two sensing electrodes.[3]. A CRPSD provides the centroid of the light distribution with a very fast response time (in the order of 10 MHz). Theory predicts that a CRPSD provides very precise measurement of the centroid versus a DRPSD. By precision, we mean measurement uncertainty. It depends among other things on the signal to noise ratio and the quantization noise. In practice, precision is important but accuracy is even more important. A CRPSD is in fact a good estimator of the central location of a light distribution. [1]
DISCRETE RESPONSE POSITION SENSITIVE DETECTORS
(DRPSD)
DRPSD on the other hand comprise detectors such as Charge Coupled Devices (CCD) and arrays of photodiodes equipped with a multiplexer for sequential reading. They are slower because all the photo-detectors have to be read sequentially prior to the measurement of the location of the peak of the light distribution. [1] DRPSDs are very accurate because of the knowledge of the distribution but slow. Obviously, not all photo-sensors contribute to the computation of the peak. In fact, what is required for the measurement of the light distribution peak is only a small portion of the total array. Once the pertinent light distribution (after windowing around an estimate around the peak) is available, one can compute the location of the desired peak very accurately.

3a) In a monochrome system, the incoming beam is split into two components

3b) artistic view of a smart sensor with colour capabilities,
Figure 3 shows schematically the new smart position sensor for light spot measurement in the context of 3D and colour measurement. In a monochrome range camera, a portion of the reflected radiation upon entering the system is split into two beams (Figure 3a). One portion is directed to a CRPSD that determines the location of the best window and sends that information to the DRPSD.

In order to measure colour information, a different optical element is used to split the returned beam into four components, e.g., a diffractive optical element (Figure 3b). The white zero order component is directed to the DRPSD, while the RGB 1storder components are directed onto three CRPSD, which are used for colour detection (Figure 3c). The CRPSDs are also used to find the centroid of the light distribution impinging on them and to estimate the total light intensity The centroid is computed on chip with the well-known current ratio method i.e. (I1-I2)/ (I1+I2) where I1 and I2 are the currents generated by that type of sensor. The weighed centroid value is fed to a control unit that will select a sub-set (window) of contiguous photo-detectors on the DRPSD. That sub-set is located around the estimate of the centroid supplied by the CRPSD.Then, the best algorithms for peak extraction can be applied to the portion of interest.

PROPOSED SENSOR “COLORANGE
ARCHITECTURE
Figure5 shows the proposed architecture for the colorange chip.
Figure5 The proposed architecture for the colorange chip
An object is illuminated by a collimated RGB laser spot and a portion of the reflected radiation upon entering the system is split into four components by a diffracting optical element as shown in figure 4b. The white zero order component is directed to the DRPSD, while the RGB 1storder components are directed onto three CRPSD, which are used for colour detection (Figure 3c). The CRPSDs are also used to find the centroid of the light distribution impinging on them and to estimate the total light intensity The centroid is computed on chip with the well-known current ratio method i.e. (I1-I2)/ (I1+I2) where I1 and I2 are the currents generated by that type of sensor. [3] The weighed centroid value is fed to a control unit that will select a sub-set (window) of contiguous photo-detectors on the DRPSD. That sub-set is located around the estimate of the centroid supplied by the CRPSD. Then, the best algorithms for peak extraction can be applied to the portion of interest.
32 PIXEL PROTOTYPE CHIP
Figure6 shows the architecture and preliminary experimental results of a first prototype chip of a DRPSD with selectable readout window. This is the first block of a more complex chip that will include all the components illustrated in Figure 3c.

The prototype chip consists of an array of 32 pixels with related readout channels and has been fabricated using a 0.mm commercial CMOS process. The novelties implemented consist in a variable gain of the readout channels and a selectable readout window of 16 contiguous pixels. Both features are necessary to comply with the requirements of 3D single laser spot sensors, i.e., a linear dynamic range of at least 12 bits and a high 3D data throughput. In the prototype, many of the signals, which, in the final system are supposed to be generated by the CRPSDs, are now generated by means of external circuitry. The large dimensions of the pixel are required, on one side to cope with speckle noise and, on the other side, to facilitate system alignment. Each pixel is provided with its own readout channel for parallel reading. The channel contains a charge amplifier and a correlated double sampling circuit (CDS). To span 12 bits of dynamic range, the integrating capacitor can assume five different values. In the prototype chip, the proper integrating capacitor value is externally selected by means of external switches C0-C4 . In the final sensor, however, the proper value will be automatically set by an on chip circuitry on the basis of the total light intensity as calculated by the CRPSDs.[1]
During normal operation, all 32 pixels are first reset at their bias value and then left to integrate the light for a period of 1ms. Within this time the CRPSDs and an external processing unit estimate both the spot position and its total intensity and those parameters are fed to the window selection logic. After that, 16 contiguous pixels, as addressed by the window selection logic, are read out in 5ms, for a total frame rate of 6ms. Future sensors will operate at full speed, i.e. an order of magnitude faster. The window selection logic, LOGIC_A, receives the address of the central pixel of the 16 pixels and calculates the address of the starting and ending pixel. The analog value at the output of the each CA within the addressed window is sequentially put on the bitline by a decoding logic, DECODER, and read by the video amplifier .LOGIC-A generates also synchronization and end-of-frame signals which are used from the external processing units. LOGIC_B instead is devoted to the generation of logic signals that derive both the CA and the CDS blocks. To add flexibility also the integration time can be changed by means of the external switches T0-T4.The chip has been tested and its functionality proven to be in agreement with specifications. [1]

ADVANTAGES
Reduced size and cost
Better resolution at a lower system cost
High reliability that is required for high accuracy 3D vision systems
Complete images of visible surfaces that are rather featureless to the human eye or a video camera can be generated.
.
DIS ADVANTAGE

The elimination of all stray light in an optical system requires sophisticated techniques.
APPLICATIONS
Intelligent digitizers will be capable of measuring accurately and simultaneously colour and 3D
For the development of hand “held 3D cameras
Multiresolution random access laser scanners for fast search and tracking of 3D features
FUTURE SCOPE
Anti reflecting coating film deposition and RGB filter deposition can be used to enhance sensitivity and for colour sensing
CONCLUSION
The results obtained so far have shown that optical sensors have reached a high level of development and reliability those are suited for high accuracy 3D vision systems. The availability of standard fabrication technologies and the acquired know-how in the design techniques, allow the implementation of optical sensors that are application specific: Opto-ASICs. The trend shows that the use of the low cost CMOS technology leads competitive optical sensors. Furthermore post-processing modules, as for example anti reflecting coating film deposition and RGB filter deposition to enhance sensitivity and for colour sensing, are at the final certification stage and will soon be available in standard fabrication technologies. The work on the Colorange is being finalized and work has started on a new improved architecture.
REFERENCES
[1] L.GONZO, A.SIMONI, A.GOTTARDI, DAVID STAPPA, J.A BERNALDIN,Sensors optimized for 3D digitization, IEEE transactions on instrumentation and measurement, vol 52, no.3, June 2003, pp.903-908.
[2] P.SCHAFER, R.D.WILLIAMS. G.K DAVIS, ROBERT A. ROSS, Accuracy of position detection using a position sensitive detector, IEEE transactions on instrumentation and measurement, vol 47, no.4, August 1998, pp.914-918
[3] J.A.BERALDIN,Design of Bessel type pre-amplifiers for lateral effect photodiodes, International Journal Electronics, vol 67, pp 591-615, 1989.
[4] A.M.DHAKE,Television and video engineering, Tata Mc Graw Hill
[5] X.ARREGUIT, F.A.VAN SCHAIK, F.V.BAUDUIN, M.BIDIVILLE, E.RAEBER,A CMOS motion detector system for pointing devices, IEEE journal solid state circuits, vol 31, pp.1916-1921, Dec 1996
[6] P.AUBERT, H.J.OGUEY, R.VUILLEUNEIR, Monolithic optical position encoder with on-chipphotodiodes,IEEE J.solid state circuits, vol 23, pp.465-473, April 1988
[7] K.THYAGARAJAN,A.K.GHATAK,Lasers-Theory and applications, Plenum Press
[8] N.H.E.Weste, Kamran Eshraghian,Principles of CMOS VLSIdesign, A systems perspective, Low Price Edition

CONTENTS
INTRODUCTION
COLOUR 3-D IMAGING TECHNOLOGY
LASER SENSORS FOR 3-D IMAGING
POSITION SENSITIVE DETECTORS
PROPOSED SENSOR-COLORANGE
ADVANTAGES
DISADVANTAGES
APPLICATIONS
FUTURE SCOPE
CONCLUSION
REFERENCES

Reply
#2
I need General report on 3D Digitization, What is 3D Digitization and then i would like to know more about the sensors on 3D Digitization, Please someone help me.. its urgent
Reply
#3
to get better idea and report then ppt of Sensors on 3D digitalization please read following page (most of these page added report)
http://studentbank.in/report-sensors-on-...ull-report
http://studentbank.in/report-sensors-on-...tion--7100
http://studentbank.in/report-sensors-on-...tion--2059
Reply
#4
full documentation report of sensors on 3D digitization

[/font][font=Times New Roman]
Reply
#5
to get information about the topic"sensors on 3d digitization" full report ppt and related topic refer the page link bellow

http://studentbank.in/report-sensors-on-...tion--2059

http://studentbank.in/report-sensors-on-...ull-report

http://studentbank.in/report-sensors-on-3d-digitization

http://studentbank.in/report-sensors-on-...tion--7100

Reply

Important Note..!

If you are not satisfied with above reply ,..Please

ASK HERE

So that we will collect data for you and will made reply to the request....OR try below "QUICK REPLY" box to add a reply to this page
Popular Searches: sensors optimized for 3d digitization ieee pdf, sensors 3d digitization seminar report, why sensors are used for 3d digitization, sensors on 3d digitization, triangulation technique in sensors on 3d digitization, model of trace passing alarmsensors on 3d digitization pdf, seminar report on sensors on 3d digitalization,

[-]
Quick Reply
Message
Type your reply to this message here.

Image Verification
Please enter the text contained within the image into the text box below it. This process is used to prevent automated spam bots.
Image Verification
(case insensitive)

Possibly Related Threads...
Thread Author Replies Views Last Post
  Transparent electronics full report seminar surveyer 8 24,352 04-04-2018, 07:54 AM
Last Post: Kalyani Wadkar
  wireless charging through microwaves full report project report tiger 90 70,541 27-09-2016, 04:16 AM
Last Post: The icon
  Wireless Power Transmission via Solar Power Satellite full report project topics 32 50,214 30-03-2016, 03:27 PM
Last Post: dhanabhagya
  surge current protection using superconductors full report computer science technology 13 26,845 16-03-2016, 12:03 AM
Last Post: computer science crazy
  paper battery full report project report tiger 57 61,682 16-02-2016, 11:42 AM
Last Post: Guest
  IMOD-Interferometric modulator full report seminar presentation 3 11,361 18-07-2015, 10:14 AM
Last Post: [email protected]
  SEASON BASED STREET LIGHT SWITCHING BASED ON SENSORS project report helper 2 3,937 29-04-2015, 01:47 PM
Last Post: seminar report asees
  digital jewellery full report project report tiger 36 66,485 27-04-2015, 01:29 PM
Last Post: seminar report asees
  Seminar Report On Wearable Bio-Sensors mechanical wiki 3 5,694 30-03-2015, 10:07 AM
Last Post: seminar report asees
  LOW POWER VLSI On CMOS full report project report tiger 15 22,191 09-12-2014, 06:31 PM
Last Post: seminar report asees

Forum Jump: