Choosing the right hardware
Many options exist when deciding upon the hardware that will be running your machine vision AI application. Field programmable gate arrays (FPGAs), graphics processing units (GPUs) and even microcontrollers (MCUs) each have their own benefits.
FPGAs are very powerful processing units that can be configured to meet the requirements of almost any application. Tailored-made FPGA architectures can be created for handling specific applications. FPGA achieves much higher performance, lower costs and better power efficiency compared to other options like GPUs and CPUs. GPUs are specialised processors that are mainly designed to process images and videos. Compared to CPUs, they are based on simpler processing units but host a much larger number of cores. This makes GPUs excellent for applications in which large amounts of data need to be processed in parallel, such as image pixels or video codecs. CPUs have a limited core count, which inhibits their ability to quickly process the large amounts of data needed for AI.
Image sensor and lighting
When developing a machine vision system, selecting the right image sensor could be one of the most important design decisions. The design requires high-resolution image capture, fast data transfer with minimal noise, and efficient processing power which is able to crunch data and generate outputs. The advancements in front-side (FSI) and back-side (BSI) illumination in CMOS sensor technology allow for higher-resolution images in low light.
Proper lighting is also an important consideration. The basis for all lighting performance comes down to three main image sensor characteristics: quantum efficiency (QE), dark current and saturation capacity. When implemented within a camera, the maximum QE of the camera should be less than that of the sensor, due to external optical and electronic effects.
Dark current and saturation capacity are also important design considerations in Machine vision systems. Dark current measures the variation in the number of electrons that are thermally generated within the CMOS imager and can add noise. Saturation capacity denotes the number of electrons that an individual pixel can store. They can be used with QE measurements to derive maximum signal-to-noise ratio (S/N), absolute sensitivity and the dynamic range of an application.
The right lighting will help increase the accuracy and efficiency of a machine vision application. Other factors to consider with lighting include wavelength, such as infrared, fixed lighting, and even lighting placement. Light sources and reflections that shine directly into the cameras of Machine vision systems have been shown to decrease object detection accuracy.
Choosing the right machine vision camera
Recent advancements in machine vision technology now let cameras transfer high-megapixel images at extremely fast frame rates. Selecting the best interface requires a review of several considerations such as Choosing a Sensor Type (CMOS or CCD), Color Camera or Monochrome Camera, Camera Output Format (GigE, Camera Link, CoaXPress, USB3, HD-SDI), and Frame Rates. CCDs have higher image quality, better light sensitivity, greater noise performance, and an ideal global shutter. CMOS sensors are known for their high speed, on-chip system integration, and low cost of manufacturing.
Camera manufacturers leverage the latest sensor developments and improvements in camera design, helping machine vision system developers and integrators create faster, more flexible, and more capable imaging systems. With higher camera resolutions comes the need for higher-quality, larger-format optics, which are readily available, with options including embedded liquid lenses for auto-focusing systems. Optics for nonvisible wavelengths enable new ways to detect things with specialised imaging using wavelengths that range from the UV through to the IR bands.
LED illumination products, critical to all machine vision applications, now come in a wide variety of wavelengths and form factors. They feature increased flexibility, with tunable angles and additional wavelengths, more consistent spectral response, and even programmable sources with embedded controls. An important enabler is the emergence of up to 100 G interfaces as well as the recently updated CoaXPress 2.0 interface and even PCI interfaces.
Picking a machine vision lens
Deciding on the right lens for a machine vision application calls for a review of the required specs, some math, and a consideration of how the lens will integrate with the camera setup. When choosing the lens used in a machine vision application, one must consider the sensor that will be used. Sensor size and pixel size are of extreme importance in the selection process. The lens must be able to properly illuminate the complete sensor area to avoid shading and vignetting.
Ideal lenses produce images that perfectly match the object captured, including all details and brightness variations. Standard lenses may be about a megapixel in fixed focal lengths of 4.5-100mm. Macro lenses are optimised for close-up focus. When selecting the correct lens for an application, designers calculate the needed working distance using 3 factors: focal length, the length of the