It is the Image Signal Processing that defines the final image quality on the screen
Many end-users shell out the cash to acquire the newest high-end devices, plug in, and expect to be wowed

A well-developed surveillance system can give a single security guard the power to see what otherwise might take a hundred pairs of eyes to see. But what happens when all the components are all connected and powered up, and the resulting image on the screen is, well, indiscernible, or, at the very least, terribly pixelated?

Many end-users shell out the cash to acquire the newest high-end devices, plug in, and expect to be wowed. Often enough, however, what they see on the screen is not what they were expecting – and they wonder what they just paid for. In a good high-definition system, what factors actually create the best image quality? With so many variables involved, from the camera’s lens to the imaging algorithms to the monitor resolution – just to name the obvious ones – how do system integrators achieve the best on-screen images?   

The lens

The first component to handle light from an object, this may be the one most taken for granted in cameras of any sort. (Just try scratching or cracking one and you’ll agree.) In the days of analogue cameras, it seemed that any old lens would do just fine. However, as the technology inside cameras evolved and more powerful sensors (more pixels) became available, engineers and programmers demanded more from lenses. Moreover, intelligent video content analyses would be impossible without high-accuracy lenses.

In what way do lenses impact the image quality? The key factor here is light transmission. The quality of light passing through the lens itself will forever be critical to the quality of image reproduced. A lens made using ultra-precision molding aspherical technology achieves more accurate colour, better light, and clearer images. Multilayer broadband anti-reflection coating further maximises a lens's light transmission while minimising the residual reflection of light on the surface of each optical lens.

Actual image performance depends upon variables such as low light illumination, signal to noise ratio, dynamic range of light, and more
Variables involved include the camera’s lens, the imaging algorithms and the monitor resolution


When it comes to fabricating a megapixel lens that hits the mark, the materials used and the processes by which lenses are produced are the two most critical criteria. The materials most often used to create lenses are glass and specialised plastics. An HD lens made of ultra-low-dispersion optical glass – which, by using dispersion characteristics that are different from those of conventional optical glass – will deliver better HD performance. Machine-automated lens production using specialised plastics results in high output for camera producers, and the lenses produced are more uniform in design and quality.

For an HD vari-focal lens, its image quality depends largely on the precision of the cam. The cam rotates to drive the zoom and focus lens groups forward and backward for a smooth continuity of focal length and adjustment of the focal point. A lack of precision with the cam inevitably causes an offset or tilt of the lens' optical axis during zooming and focusing, leading to a serious loss of image quality.

Lens production is a delicate
balancing act. The slightest
errors or imperfections will
be very noticeable when tested

Lens production is a delicate balancing act. The slightest errors or imperfections will be very noticeable when tested. The features of a lens that affect image resolution, clarity, and contrast must be perfect. Achieving uniformity of image resolution at the centre and the edges of a lens requires high-precision machinery. And once a lens has been properly crafted, the assembly of the camera, the lens housing materials, and the alignment of the optical axis demand utmost accuracy. To put it mildly, quality control must be rigorous.

Image signal processing

As light passes through the lens, the sensor captures it and converts it to data. Raw RGB data is transmitted by the camera sensor and undergoes Image Signal Processing (ISP) such as noise reduction, white balance, WDR, curve correction and colour correction, etc. The data is then transformed to true colours for each pixel point, for people to see images that look “normal” to the human eye. It is the Image Signal Processing that defines the final image quality on the screen.

Collecting data in different conditions is vital, for instance, outdoor data should be analysed with natural light on days with sun, overcast, rain, and fog, at dawn, at dusk, and so on. Similarly, when using cameras equipped with infrared sensors, testing the IR light signals in various conditions is necessary as well. 

Actual image performance depends upon variables such as low light illumination, signal to noise ratio, dynamic range of light, and more. ISP algorithms aim at increasing the signal data and decreasing noise. Cameras with Wide Dynamic Range (WDR) will yield improved video imaging with both background and foreground objects in high contrast or high-backlight environments, maximising the amount of detail in brighter and darker areas in one field-of-view. In scenes with low contrast and low light, the sensors deliver digital image signals and at the same time send some amount of digital noise that directly hinders image clarity. Three-dimensional digital noise reduction (3D DNR) removes unwanted artifacts from an image, reducing graininess. Where cloudy weather poses a challenge, auto-defogging technology helps to identify the density of fog or rain with gray-white colour ratio analysis, and imbues images with true colour reproduction.

A more efficient video encoding solution would allow an improvement in compression efficiency of 40–50% over H.264
Ramping up the megapixels and frame rates yields great video, but also results in more bandwidth used and more storage occupied


Matching megapixels to image quality

When the factors mentioned above line up well, correlating cameras and monitors creates the best viewing experience. When a high definition camera is in place, a monitor with a high resolution will display images much more clearly. But if the monitor’s resolution is low, it will not deliver the high-quality images expected – or possible – from that HD camera. For an 8 MP camera, for instance, users do best to apply monitors with 4K × 2K resolution. Though common sense, this deserves to be mentioned because users might decide to upgrade their systems with 4K monitors, but with perhaps 1.3 MP cameras installed. In such a scenario, there’s no guarantee the on-screen image quality will automatically improve.

Managing data and bandwidth

In terms of a complete, high definition surveillance system, when the right factors come together and the calibrations are set, image quality – even in a standard HD 1080p setup – can be extremely good. The final piece of the puzzle is managing the data. Ramping up the megapixels and frame rates yields great video, but also results in more bandwidth used and more storage occupied. Squeezing bandwidth threatens image quality and clarity, but keeping ample room for signal transmission and storage will eventually increase the overall cost for customers. Is it possible for integrators to optimise their customer’s system and, at the same time, stay within budget constraints? Luckily, it can be done.

Squeezing bandwidth threatens
image quality and clarity, but
keeping ample room for signal
transmission and storage increases
the overall cost for customers

To do this, a more efficient video encoding solution would allow an improvement in compression efficiency of 40–50% over H.264. Improvements to algorithms that are adaptive to a particular scene give users control over bitrate. Another option would be to start recording video only when an event triggers an alarm, since most security guards are primarily concerned with moving objects rather than a scene’s generally stagnant background. This intelligently helps optimise bandwidth and storage consumption. Another method is to use a single panoramic or fisheye camera in place of several HD cameras for coverage – the reduced number of security devices will reduce bandwidth demands and the rate of storage consumption as well.

Getting the best image quality

Now let’s put this all together. Naturally, integrators and users will refer to their product specs to understand features and functions, fine-tuning each component for best results. Also, as suggested above, users should select an HD camera comprehensively in terms of lens performance, pixels, image quality, and overall system compatibility and performance. Next, matching the backend device and management platform should be carefully considered in a complete security system. Installing equipment that has been engineered for a given scene is a must, along with strategising how to get the most coverage out of the lowest number of cameras. Finally, product quality, warranty, price, and on-going customer service are all important factors that customers should take into account as well.

Download PDF version

Author profile

Can You Senior Embedded Software Development Manager, Hikvision

In case you missed it

What is the changing role of training in the security industry?
What is the changing role of training in the security industry?

Even the most advanced and sophisticated security systems are limited in their effectiveness by a factor that is common to all systems – the human factor. How effectively integrators install systems and how productively users interface with their systems both depend largely on how well individual people are trained. We asked this week’s Expert Panel Roundtable: What is the changing role of training in the security and video surveillance market?

What is AI Face Search? Benefits over facial recognition systems
What is AI Face Search? Benefits over facial recognition systems

When a child goes missing in a large, crowded mall, we have a panicking mom asking for help from the staff, at least a dozen cameras in the area, and assuming the child has gone missing for only 15 minutes, about 3 hours’ worth of video to look through to find the child. Typical security staff response would be to monitor the video wall while reviewing the footage and making a verbal announcement throughout the mall so the staff can keep an eye out for her. There is no telling how long it will take, while every second feels like hours under pressure. As more time passes, the possible areas where the child can be will widen, it becomes more time-consuming to search manually, and the likelihood of finding the child decreases. What if we can avoid all of that and directly search for that particular girl in less than 1 second? Artificial neural networks are improving every day and now enable us to search for a person across all selected camera streamsWith Artificial Intelligence, we can. Artificial neural networks are improving every day and now enable us to search for a person across all selected camera streams in a fraction of a second, using only one photo of that person. The photo does not even have to be a full frontal, passport-type mugshot; it can be a selfie image of the person at a party, as long as the face is there, the AI can find her and match her face with the hundreds or thousands of faces in the locations of interest. The search result is obtained in nearly real time as she passes by a certain camera. Distinguishing humans from animals and statues The AI system continuously analyses video streams from the surveillance cameras in its network, distinguishes human faces from non-human objects such as statues and animals, and much like a human brain, stores information about those faces in its memory, a mental image of the facial features so to speak. When we, the system user, upload an image of the person of interest to the AI system, the AI detects the face(s) in that image along with their particular features, search its memory for similar faces, and shows us where and when the person has appeared. We are in control of selecting the time period (up to days) and place (cameras) to search, and we can adjust the similarity level, i.e., how much a face matches the uploaded photo, to expand or fine-tune the search result according to our need. Furthermore, because the camera names and time stamps are available, the system can be linked with maps to track and predict the path of the person of interest. AI Face Search is not Face Recognition for two reasons: it protects people’s privacy, and it is lightweight Protecting people’s privacy with AI Face Search  All features of face recognition can be enabled by the system user, such as to notify staff members when a person of interest is approaching the store AI Face Search is not Face Recognition for two reasons: it protects people’s privacy, and it is lightweight. First, with AI Face Search, no names, ID, personal information, or lists of any type are required to be saved in the system. The uploaded image can be erased from the system after use, there is no face database, and all faces in the camera live view can be blurred out post-processing to guarantee GDPR compliance. Second, the lack of a required face database, a live view with frames drawn around the detected faces and constant face matching in the background also significantly reduces the amount of computing resource to process the video stream, hence the lightweight. Face Search versus Face Recognition AI Face Search Face Recognition Quick search for a particular person in video footage Identify everyone in video footage Match detected face(s) in video stream to target face(s) in an uploaded image Match detected face(s) in video stream to a database Do not store faces and names in a database Must have a database with ID info Automatically protect privacy for GDPR compliance in public places May require additional paperwork to comply with privacy regulations Lightweight solution Complex solution for large-scale deployment Main use: locate persons of interest in a large area Main use: identify a person who passes through a checkpoint Of course, all features of face recognition can be enabled by the system user if necessary, such as to notify staff members when a person of interest is approaching the store, but the flexibility to not have such features and to use the search tool as a simple Google-like device particularly for people and images is the advantage of AI Face Search.Because Face Search is not based on face recognition, no faces and name identifications are stored Advantages of AI Face Search Artificial Intelligence has advanced so far in the past few years that its facial understanding capability is equivalent to that of a human. The AI will recognise the person of interest whether he has glasses, wears a hat, is drinking water, or is at an angle away from the camera. In summary, the advantages of Face Search: High efficiency: a target person can be located within a few seconds, which enables fast response time. High performance: high accuracy in a large database and stable performance, much like Google search for text-based queries. Easy setup and usage: AI appliance with the built-in face search engine can be customised to integrate to any existing NVR/VMS/camera system or as a standalone unit depending on the customer’s needs. The simple-to-use interface requires minimal training and no special programming skills. High-cost saving: the time saving and ease of use translate to orders of magnitude less manual effort than traditionally required, which means money saving. Scalability: AI can scale much faster and at a wider scope than human effort. AI performance simply relies on computing resource, and each Face Search appliance typically comes with the optimal hardware for any system size depending on the customer need, which can go up to thousands of cameras. Privacy: AI Face Search is not face recognition. For face recognition, there are privacy laws that limits the usage. Because Face Search is not based on face recognition, no faces and name identifications are stored, so Face Search can be used in many public environments to identify faces against past and real-time video recordings. AI Face Search match detected face(s) in video stream to target face(s) in an uploaded image Common use cases of AI Face Search In addition to the scenario of missing child in a shopping mall, other common use cases for the AI Face Search technology include: Retail management: Search, detect and locate VIP guests in hotels, shopping centres, resorts, etc. to promptly attend to their needs, track their behaviour pattern, and predict locations that they tend to visit. Crime suspect: Quickly search for and prove/disprove the presence of suspects (thief, robber, terrorist, etc.) in an incident at certain locations and time. School campus protection: With the recent increase in number of mass shootings in school campuses, there is a need to identify, locate and stop a weapon carrier on campus as soon as possible before he can start shooting. Face Search will enable the authorities to locate the suspect and trace his movements within seconds using multiple camera feeds from different areas on campus. Only one clear image of the suspect’s face is sufficient. In the race of technology development in response to business needs and security concerns, AI Face Search is a simple, lightweight solution for airports, shopping centres, schools, resorts, etc. to increase our efficiency, minimise manual effort in searching for people when incidents occur on site, and actively prevent potential incidents from occurring. By Paul Sun, CEO of IronYun, and Mai Truong, Marketing Manager of IronYun

What technology will impact security most in the rest of 2018?
What technology will impact security most in the rest of 2018?

Where does the time go? Before you know it, here we are at mid-year reflecting on an eventful first half of 2018 in the physical security market. It’s also a good time for our Expert Panel Roundtable to pause and look ahead at what we might expect in the second half of the year. We asked this week’s Expert Panel Roundtable: What technology development will have the greatest impact in the second half of 2018?