Rotronic is pleased to announce the introduction of its smallest ever temperature and humidity logger. The HL-1D measures only 90 x 60 x 23 mm, is well specified with good accuracy, durable and has high ingress protection against dust and water (IP67). HW4-Lite validated software for programming, data download and analysis is included. The logger is available now at a competitive, inexpensive price.
The Rotronic HL-1D logger is very suitable for monitoring and recording conditions for a wide range of applications across all industries, in commerce and for research organisations. The compact logger is particularly suitable for monitoring high value products of all types during transportation to ensure that quality is maintained.
The significant features of the HL-1D inexpensive logger include:
• High measurement accuracy of 3 %rh and 0.3 °C
• Slim, compact and durable. High ingress protection rating of IP67
• Clear LC display with configurable visual alarms
• Large data recording capacity: 32,000 data points
• Ranges: -20…70 °C, 0…100 %rh. Logging interval from 30 s
• Min / Max / Average statistical function on logger display
• Package includes HW4-Lite validated PC software with data analysis
• Three year battery life (with five minute logging interval)
• FDA 21CFR Part 11 / GAMP5 conformity
• Temperature only logger available (product code TL-1D)
HL-1D / TL-1D Logger Technical Datasheet – click here
Contact us now for logger pricing and additional information
Companies across many industries needing to perform regular monitoring and calibration have never faced a more challenging environment. Stricter compliance requirements mean companies are under greater pressure to deliver accurate and reliable data, whilst internal budget restrictions demand the most cost effective and efficient solutions.
Can modern measurement & calibration techniques help your business operations?
It is well known that accurate measurements reduce energy use and improve product consistency. Instrument users, calibration laboratories and manufacturers are constantly looking for smarter ways of operating and are responding with innovations that are transforming the measurement and calibration industry.
New ways of working
Industrial environments are now more automated and interconnected than ever before and companies need to ensure that their infrastructure and processes have the ability to respond and adapt to industry changes. With the introduction of newer, more complex instrumentation, organisations can often be slow to recognise the additional business benefits that can be achieved by replacing a traditional method that (offers a short term result) with a more modern method (that delivers a longer term sustainable solution). Implementing a new approach can also help re-position the calibration process from being viewed simply as a cost to business to one that helps deliver improved process and energy efficiencies with a return on investment.
Historically, in-situ calibration has been the standard approach; however, advances in technology means that there is now a viable alternative whilst still maintaining the growing demand for on-site services. With the market moving away from analogue to digital signal processing, interchangeable digital sensors are proving to be a more practical solution for both large and small organisations alike. As businesses look for greater automation and productivity, modern interchangeable digital sensors are allowing calibration to be achieved much more quickly without the costly implications of operational downtime and on-site maintenance.
Why calibrate? – The only way to confirm performance In unsettled economic times it can be tempting to simply extend the intervals between calibration cycles or to forgo calibration altogether. However, neglecting system maintenance and calibration will result in reduced performance and a loss of measurement confidence, ultimately leading to a failure to meet compliance standards. Measurement drift over time negatively impacts on processes and quality. Regular, accredited calibration demonstrates compliance, but equally importantly, sends a message to customers that quality is taken seriously and that they can be confident in both the process and the final product.
Traditional In-Situ Sensor Calibration
Until recently most humidity calibrations were performed on-site in-situ. Larger organisations with multiple instruments generally found it more convenient to have their own in-house calibration instruments with dedicated technicians working on-site. Smaller organisations unwilling or unable to invest in on-site calibration equipment had the option to engage the services of a commercial calibration provider.
In most cases, trained instrument technicians are required for in-situ calibration work; the equipment is brought to the probes and generally only one probe can be calibrated at a time. One of the main disadvantages of this process is the impact that it has on production downtime, as typically a salt or chamber based calibration can take more than three hours. Moreover, as the processes or control systems are interrupted during calibration, the actual conditions can be unknown.
Modern Ex-Situ Sensor Calibration
Companies keen to avoid the impacts of in-situ calibration and/or operational downtime caused by the replacement of failed hard wired instruments are opting instead for the flexibility and convenience of interchangeable sensors and modern portable calibration generators. Instead of bringing in equipment to calibrate in-situ, the technician brings pre-calibrated probes directly from the laboratory (on-site or external). Using interchangeable digital sensors, the pre-calibrated probes can be exchanged with the in-situ probes in seconds (known as hot swaps), saving time and avoiding operational disruption. If a wider system loop calibration is required, digital simulators are applied and provide any fixed values exactly and instantly. The old probes are then taken back to a calibration laboratory and calibrated accordingly. This adds the benefit that an external accredited laboratory can be used without issue.
Improved accuracy and traceability?
By ensuring that all calibrations are performed within dedicated laboratories as opposed to ad-hoc locations, better procedures and instrumentation can be utilised. In addition, time pressures are usually reduced as processes and monitoring systems are unaffected during calibration. As such calibrations are typically performed to a higher standard leading to lower associated measurement uncertainty (every calibration will have an uncertainty associated with it – whether it is defined or not). Overall in most circumstances these methods deliver greater reliability, improved traceability and importantly, reduces on-site workload and limits operational downtime.
CASE STUDY – Meeting the demands at the National Physical Laboratory, London.
When the National Physical Laboratory (NPL) in London needed to replace their entire building management system (BMS), they turned to Rotronic Instruments (UK) for an integrated solution to the sensors and calibration. The NPL was looking for both a complete range of temperature and humidity sensors and instrumentation, and the fulfilment of the calibration and commissioning needs of these instruments. Working closely with the project stakeholders, the Rotronic Instruments (UK) team developed a tailored solution, matching the instruments and service to the project requirements.
The decision by the NPL to replace the BMS was brought about by the need for tighter control, greater reliability and easier calibration. One of the key elements in achieving these objectives was the use of interchangeable probes. This immediately limited time-consuming and disruptive on-site sensor calibration to a minimum. Every probe’s digital output was calibrated in Rotronic Instruments’ (UK) UKAS accredited laboratory, and each transmitter’s analogue output was calibrated using a simulated digital input. To resolve any measurement errors in-situ between the calibrated sensors and uncalibrated BMS, each installed high accuracy instrument was loop calibrated and adjusted. Typical installations errors corrected for to date on the brand new BMS are ±0.5 %rh and ±0.25°C; a significant result for labs requiring tolerances of better than 1 %rh and 0.1°C.
Whilst the use of high performance instruments was essential, not every sensor location or application could justify this approach. However, mindful of the NPL’s long term objectives, even the lowest specification thermistor products were customised to provide long-term performance and low drift. Additionally, a robust commissioning procedure and training for key personnel was developed to enable ongoing commitment to delivering quality measurements. Finally, it was effective communication and regular on-site interaction with all the stakeholders that helped deliver a successful outcome to this substantial project.
All companies that need to perform regular monitoring and instrument calibration should be constantly reviewing their processes and questioning whether their operations and procedures are delivering the maximum return for their business. As increased regulatory compliance and demands for improved energy efficiencies continue to grow, traditional processes may no longer offer the optimum solution. An organisational mindset change may be needed to move calibration from being seen as a fixed cost to a process that can help deliver business objectives through ongoing cost and energy efficiencies.
With the advent of calibration methods that can significantly reduce in-situ disruption, downtime is minimised, labour costs are reduced and productivity improved. Using interchangeable digital systems increases the accuracy and traceability of calibrations, resulting in higher quality product.
Choosing the right calibration methodology may require new thinking and a different approach, but those companies that get it right will end up with a modern, flexible system that both achieves compliance and delivers long term cost and energy efficiencies to their business.
For more information on the NPL case study or how your business can develop innovative and efficient monitoring solutions please contact us.
As we continue to measure relative humidity in more and more environments with ever increasing accuracy demands, we are pushing the humble capacitive humidity sensor into new realms.
Accuracy, drift, operating range and chemical resistance are key challenges for the relative humidity sensor industry. Our sensor experts work hard to develop new polymers and construction methods to ensure the best performance. At the same time advanced electronics and probe housings enable digital calibration and complex temperature corrections to further increase accuracy and performance. A final and often neglected part of ensuring a relative humidity probes performance is its filter. The correct filter ensures fast response and environmental protection. Filters also offer mechanical protection and eliminate damage caused by extreme airflow.
However understanding why sensors fail is often difficult to predict or understand. In many cases the chemicals and contaminants that sensors are exposed to are unknown. In these situations often selecting the best sensor can only be achieved through mutual relationships built around quality support and service.
In the UK we have worked closely with many customers and in combination with our Swiss technical divisions to select and develop solutions for some highly aggressive and challenging environments. Some of these projects are examined below in more detail.
Hydrogen peroxide vapour sterilisation.
– Hydrogen peroxide vapour is used to chemically sterilise environments and products by generating a vapour of toxic Hydrogen Peroxide. When the humidity reaches the dew point of the surfaces condensation forms sterilising all surfaces. However the chemicals are also highly aggressive to humidity sensors and constant cycles of saturation worsen the effects.
– Making use of Rotronic’s specifically designed H2O2 resistant sensor as well as additional conformal coating to protect exposed connections in further combination with improved customer understanding around handling and storage, has resulted in a solution that has exceeded customer expectations. Importantly, whilst this was not achieved first time around, through a partnership driven towards the end goal we achieved success.
Chemical degradation on the sensor surface
– Accelerated commercial composting is an impressive sight to see. The chemical and biological processes occurring are complex and surprisingly aggressive. The wrong materials can literally become part of the final compost if you are not careful. Chemically resistant sensors help to provide some longevity to instruments but one of the key areas requiring extra attention is around cable and filter design. Modifying a standard industrial grade sensor with bio-resilient cables ensures the probes are not eaten alive!
Highly accelerated life testing.
– As a supplier to many chamber manufacturers and companies providing testing services this is a common application. Chambers are cycled between high and low temperatures and humidities to simulate many years aging over a short period of time. The same effects are happening to the humidity sensor – critical for the control or validation of the chamber conditions. Using industrial sensors with electronics isolated away from chamber space reduces the effects of the sudden changes. But also care taken placing the sensor away from humidity outlets and well into the chamber to avoid stem conduction all help to avoid the sensor becoming saturated as temperature cycle – which is one of the main causes for corrosion and drift. Finally, careful filter maintenance is always important.
Swimming pool monitoring and control.
– Our featured image shows chemical formation on a non-Rotronic sensors connections. Rotronic uses inert metals in the sensor design to reduce the re-activity of the sensor to chemicals in the environments. Swimming pools have a mix of high humidity, chemicals and high temperatures which work together to corrode unprotected electronics. Sensor location is key to avoid direct exposure to spray and neat chemicals. Suitable filters and if required chemical resistant sensors have proven highly successful where other instruments have failed.
So you can see not all applications are easy and we have not even begun to explore the basic issues of accurate measurement and control present with every humidity sensor installation. However our belief and aim is that through communication and partnerships we can provide the right product to ensure the desired mix of performance, resilience and price for our customers – it’s not easy but it makes life interesting!
At Rotronic UK our UKAS laboratory have worked hard to make a name for itself in high quality calibrations and service. Thanks to constant improvements in measurement procedures the laboratory is growing into one of the most advanced commercial facilities in this specialised field. The ISO 17025 accredited calibration of humidity and temperature sensors and dew point instruments confirms performance and is increasingly a requirement of industry regulations and company quality management systems. The UKAS laboratory at Rotronic UK has spent the last two years increasing confidence in the calibrations performed and as a consequence lowering the Calibration and Measurement Capability (CMC) of the laboratory. Significant improvements have been made in the measurement procedures for dew point and temperature in air, enabling the following UKAS Accredited CMCs:
Dew/Frost point measurement (°Cdp/fp) • -60 to -40 °Cfp; uncertainty ±0.14 °Cfp • -40 °Cfp to +60 °Cdp; uncertainty ±0.11 °Cdp • +60 °Cdp to 70 °Cdp; uncertainty ±0.12 °Cdp
Temperature in air/ °C • -60 °C to 0 °C; uncertainty ±0.08 °C to 0.06 °C • 0 °C to +70 °C; uncertainty ±0.05 °C • +70 °C to +150 °C; uncertainty ±0.07 °C to 0.16 °C
Relative Humidity (RH)/%rh In the laboratory RH is derived from vapour pressure formulations. Improvements in dew point and temperature in air CMCs therefore affect the RH CMCs profoundly. The improvement lies in the range 0 to 70 °C; in the worst case RH CMC is ±1.0 %rh. In all parts of the HC2-S specified range covered by the accreditation the CMC is better than the specification of the probe. This is the first time this has been achieved. With the new temperature in air calibration range (-60 °C to +150 °C) and new dew/frost point calibration range (-60 °Cdp and up to +70 °Cdp) the laboratory’s RH calibration range has been extended up to 70 °C and 98 %rh. For example, at the new upper limit of 70 °C/98 %rh the CMC is ±0.6%rh and with these levels of calibration measurement uncertainty and range of accredited calibration services the purpose-built laboratory is one of the most advanced commercial facilities in the world.
Dr. Jeremy Wingate Rotronic UK
News, views and interesting things from the world of metrology