jump to navigation

Virtual Reality – Real Insight August 8, 2017

Posted by Jon Ward in Advertising, eye tracking, Market Research, Shopper Research, Technology, Updates.
add a comment

Imagine the excitement, akin to a Christmas morning as the UPS delivery driver gets you to sign for the box that contains a thing of beauty, your shiny new VR headset. But this is no ordinary VR headset, oh no – not for you – master of all your survey, keeper of the technology, King or Queen of the early adopters! This is a VR headset with a built in eye tracker, and you are ready….. ready to dive into far off lands, explore supermarkets on the other side of the world and measure your virtual driving skills around the race track and then take over the world and….. hang on… wait a minute…. what are these instructions saying? “download the SDK and API and integrate it into your own platform….” but….. but….I don’t have time for this I have worlds to dominate! We hear you and we have the solution to your problem.

The current crop of eye tracking integrations into VR supply you with an API and SDK, and maybe some example code to stream and save data into Unity, WorldViz or similar… but not much else, and while some of the coders out there get excited about this blank canvas our experience is that most people want something they can work with pretty much right away. But it isn’t quite that easy.

Unlike screen based or glasses based eye tracking software where you have a lot of known parameters (the website is on screen and it is http://www.acuity-ets.com, or I know the user was shopping in Tesco in Croydon and this is what the shelf looked like), with VR you could be flying a spaceship down a canyon on Mars, or walking through a prototype Apple store in Singapore via a brief trip to a coffee shop to take part in a social science project. The boundless flexibility of virtual reality presents a challenge, it is almost impossible to create a platform that covers everything anyone might want to do in VR. ALMOST impossible.

At Acuity we have over 10 years’ experience in using eye tracking at the coal face, whether that is selling systems, running research, troubleshooting and advising on R&D or helping people go beyond the heatmap and get real value from the data and we know that people aren’t going to be happy with just an SDK for their new VR headset – they want more. So we made AcuityVR.

AcuityVR is a module designed to work with new or existing Unity assets, meaning that your existing investment in models can be reused, or you can build new ones and we have some amazing content partners we can recommend, if you don’t have in-house capabilities. We don’t ask you to change the way you do things, so if you use controllers for interaction then that’s fine; if you have enough space for people to walk your VR environment – not a problem for AcuityVR. The way we have designed it means that you can drop the code into a driving simulator, a walkthrough of a train station, a virtual pet grooming salon, a retail store, a clinical simulation or…. well just about anything!

We wanted to make the product a logical step from current eye tracking platforms on the market so you will see all the usual functionality in place – you can live view sessions and see the eye gaze as your participant moves around the environment; we have gaze replays, heat maps, opacity maps and statistical analysis of areas of interest – and all of these for single or multiple users. But VR gives us so much more opportunity – how about replaying multiple users at the same time while you view the environment from different angles and when you notice interesting behaviour simply switch to that persons point of view? Add multiple camera views to give different vantage points throughout the journey and view behaviour from 1st or 3rd person perspectives. We also measure and track footfall, dwell time and direction of journey – ideal for wayfinding and layout planning. Of course, being Acuity we know that eye-tracking data requires some specialised algorithms to turn data-points into fixations, and so we took care of that too, so you don’t need a PhD in eye-movements to make sense of the data.

One of the key benefits of research in VR is that you control the entire environment and this provides a number of benefits that simply can’t be matched in real world testing. Firstly every single item in an environment can be automatically classed as an AOI. That’s right, no more drawing little boxes or creating key frames. Every item in the environment can automatically have statistical data captured. In real time. If a person engaged with something you didn’t expect, no problem AcuityVR already has the stats. Measure fixations, glances, average fixations durations and time to first fixation instantly. Want to re-use the dataset for POS instead of product interaction, sure – the data is already there and rea   dy to go! All of our visualisations, replays, stats and other tools are available instantly after the session is finished – no more waiting for days for coding, number crunching or exporting image and videos. This leads to both time and cost efficiencies in research, design cycles and implementing change.

VR allows you to control the world (well, at least the virtual world – actual global domination comes later!) – imagine the possibilities of being able to dynamically change environments. Turn a daytime driving simulation into night instantly, then add some fog or a thunder storm. Take products out of stock off the shelf or change pricing in real time. Have an avatar respond negatively to the user by not making eye contact with them. Use 3D audio to totally immerse the user…. the possibilities are endless within VR, and eye tracking adds a whole new dimension for optimisation, interaction and research capabilities in virtual environments. But without a tool to capture and anlyse the data, all that potential can prove difficult to achieve. AcuityVR is that tool.

Give us a call, drop us an email or maybe send us a virtual hello at sales@acuity-ets.com or visit http://www.AcuityVR.com for more.

 

Copy Cat Brands – Who is Trying to Steal Your Attention? October 31, 2016

Posted by Jon Ward in Advertising, eye tracking, Market Research, Marketing, neuromarketing, Shopper Research.
add a comment

 

Tim from Acuity has recently been speaking at a conference in Peru where he presented some of the exciting findings from our parasitic brands research last year. Using the world leading facilities at the GSK SSL and in partnership with the British Brands Group we tested people’s recognition of famous brands and their not-so-famous imposters under a variety of conditions. Have a watch of the video below and maybe head over to the Acuity Intelligence website and read more about the study here : http://www.acuity-intelligence.com/blog/statute-of-imitations

 

3D – The Key to Tobii’s Performance Lead October 17, 2016

Posted by Scott Hodgins in eye tracking, Glasses, Market Research, Marketing, Media, neuromarketing, Shopper Research, Technology, Tips And Tricks, Tobii, Updates, Usability & UX.
Tags: , , , , , ,
add a comment

This post is trying to answer some of the most common questions that we get asked – Why should I buy a Tobii? Why is it better? System “X” has a “better head box” and system “Y” is cheaper.

The answer from our point of view is simple, the eyetracking is more accurate than using other systems for more people over a longer timeframe.

This is a pretty grand claim, why are we so confident?

Let’s start at the beginning; Eyetracking itself is straight forward, there are several well documented methods to find and follow the pupil, Tobii uses a non-intrusive video based technique called “Pupil Centre Corneal Reflection” (PCCR). Essentially an IR illuminator is used to help differentiate between the pupil and the iris, it also creates a highlight or glint that we use as well. The Tobii systems use an improved version of this idea, the secret-sauce as it were being a combination of two things, illumination and data modelling. These two areas allow the remote and wearable trackers to monitor the respondents relative 3D position in space, adjust the calibration parameters in the 3D physiological model, and therefore afford a far greater range of movement than similar systems while keeping accuracy and precision.
(Figure below shows the native 3D data from the TG2)

3d-head-coord-tobii

Illumination: Tobii can use up to two different lighting techniques known as bright and dark pupil to optimise the illumination for the participant in that location, and crucially when they move we can adapt the illumination to keep track of them. This allows a Tobii to offer people greater freedom of movement while retaining the tracking accuracy without the need for constant drift correction from the system operator.

Data modelling: The Tobii method is different having typically used multiple cameras in their research grade eyetrackers and have done since the launch of the T and X series systems in 2007/8. The advantage of using multiple cameras is that we can physically describe the location of the eye in space. That is to say we know with a very high degree of accuracy where the centre of your eye is, and which eye it is, for every sample recorded. The slightly different images from the pair of cameras in an X2 for example allows the creation of a 3D physiological model of the eyes it is tracking during calibration. This approach allows Tobii to understand the movement of the eye or the eyetracker should one or the other move and adjust the calibration accordingly with a high degree of precision.

The net result is that the these systems can accommodate movement, even if the head leaves the area trackable by the hardware and can recover tracking when the eyes are visible again, this is one of the reasons people keep choosing Tobii for demanding applications like infant research and in-vivo commercial research. In a recent study Acuity Intelligence recruited 330 people as they were entering supermarkets and didn’t have to turn away a single participant because they could not be tracked – a first for any data collection exercise with this number of people regardless of the brand of technology they were using.

Don’t just take out word for it, please challenge us, whether it is onscreen, in the real world or in the emerging AR and VR application areas we can help.

 

Neuro-Tools : Essentials September 23, 2016

Posted by eyetrackrob in Biometric, Captiv, Market Research, neuromarketing, Shopper Research, TEA, Technology.
3 comments

In recent years eyetracking has become a standard measurement in many research fields and with the “neuro”-hype many companies and universities have started to add direct and / or indirect measurements of the central nervous system to their research toolbox aiming to add an additional dimension to help understand human behaviour and decision making.

Far from being a complete catalogue of all the options currently available this series of posts will concentrate on the more practical, and commonly used, tools for commercial research – things such as salience mapping, eyetracking, facial expression analysis, electroencephalography (EEG), implicit association tests and galvanic skin response (GSR).

With the dawn of wearable fitness devices that can easily measure blood volume pulse (BVP), from which heart rate and heart rate variability (HRV) may be derived, access to these measurements have become much easier, although not without limitations as it will become clearer in this series of blogs. Additionally some of those wearable fitness devices do allow some measurement of measure electro-dermal activity (EDA) and skin temperature showing that this technology is not far from mainstream use, at least in some form.

fitness-monitor

Although the word “neuro” is very often thought as a synonym for “brain”, neuroscience comprises the study of the complete nervous system and the tools and techniques involved are suited to measure directly or indirectly certain aspects of the processes occurring within. These tools can be broadly divided into three categories : neuro measurements, behavioural measurements and biofeedback measurements. The latter is as good as a starting point as any.

Our nervous system is quite complex and can be divided into different branches which monitor and control different body functions.

One of the branches, the sympathetic nervous system (SNS), is responsible for quick fight or flight reactions. By constantly accessing the surroundings and scanning for situations that could potentially be dangerous an evaluation takes place which leads to preparations for an adequate fight or flight reaction. These preparations can be measured throughout the body and include changes in heart rate, levels of sweat on hands and feet and respiration.

sns

The reactions of the SNS are not immediate to the exposure to the stimulus to be evaluated. Reaction times and strength are highly individual and distinct for different measures. They can vary between 400 milliseconds up to 5 seconds. As part of the fight or flight reactions, the change in sweat levels on the palms and fingertips is thought to be an evolutionary mechanism allowing a firmer grip. Interestingly this reaction can also be measured on the feet!

Changes in pulse are associated with changes in either physical exercise or arousal. If physical exercise is constant, heart rate variation can be a reliable index of arousal. Research has been conducted measuring different combinations of HRV and heart rate related to stress and to the identification of positive or negative valence and even specific emotions.

A third measured physiological measurements is respiration. The perception or anticipation of odours is depended on respiration. In other words our sense of smell and therefore emotional activation through it, is enhanced by respiration. Research has associated respiration rate and depth with emotional impact and emotional valence.

tsensgsr

At Acuity we provide tools to measure biofeedback synchronized with eyetracking to help understand not only where people are looking but also the emotional impact that it is causing. We can provide a series of sensors from different manufacturers that can be brought together into Captiv L700, a software from our friends over at TEA ergo (click here to see a video of the TEA Captiv Software integrating a variety of neuro-tools).
We are also happy to help you with training to explain how those sensors work, what they are measuring and get you started on the analysis and interpretation side of things.

My next post will focus on GSR but I will cover other biometrics, EEG, facial expression analysis and complements to eyetracking data in the following posts.

Stay tuned or feel free to get in touch via sales@acuity-ets.com to learn more about how to use neuro-tools in your research.

Where Is The Value From Eyetracking August 9, 2016

Posted by Jon Ward in eye tracking, Glasses, Market Research, Marketing, Shopper Research, Technology, Tobii, Usability & UX.
add a comment

Of the past decade we have seen eyetracking move out of the research labs and academic institutes and begin to hit mainstream uses in markets such as gaming and control of operating systems, but if we look holisitically over every possible use and application for this amazing technology one question crops up quite regularly – “what can’t you eyetrack?”

Potentially this could prompt a simple answer and we list some obvious things and limitations of eyetracking as a technology – but I think the bigger question is “will eyetracking add value to what I am doing?” as it isn’t always obvious where the return on investment is from the data that eyetracking gives you.

There are actually very few situations where you can’t eyetrack people (or indeed some species of animal!) – for example recently Tobii equipment was used to eyetrack a F1 driver (https://www.youtube.com/watch?v=zjkUUMZnTnU) where the latest technology readily mounted inside the very snug and close-fitting helmet of Nico Hulkenberg. Staying with a sports theme Zoe Wimshurst from Southampton Solent University used the Tobii Glasses on a gymnast who performed a number of backflips while the equipment not only stayed in place, but also remained accurate thanks to Tobii’s 4 camera binocular platform (https://twitter.com/ZoeWimshurst/status/760472938499936256). And while we are name dropping I eyetracked Cristiano Ronaldo some years ago (https://www.youtube.com/watch?v=2NcUkvIX6no) with a previous generation platform. Sports is one thing, but what about something different… how about primate research, yup – we can tick the box there as well! For example some trials we did with Edinburgh Zoo (http://www.living-links.org/2012/11/), and the typical uses of consumer shopper and online research, psychology, linguistics and infant research are all areas where eyetracking is heavily involved, and this is of course now developing into the virtual world, mounting systems into VR and AR units for the next generation of these fields… not to mention interaction within gaming – both user testing and for control applications (http://www.tobii.com/xperience/apps/the-division/).

 

Eye tracking at Edinburgh Zoo

 

So you might say “great PR but what does that give us apart from a YouTube video” – well lets look at some examples, starting with F1 – if by watching the eye movements and point of gaze of a F1 driver we can shave 0.1 seconds per lap from a 58 lap race, we gain 6 seconds. In the Australian GP 7 drivers (from the 16 that finished) could have gained a position with this advantage, one driver could have jumped 4 places – gaining 7 points in the drivers championship in the process. For a footballer, releasing the ball 0.25 second earlier because you have the ability to read the field more efficiently visual performance training could be the difference between beating the offside trap and scoring or dropping points in a multi-billion pound race to the title. In elite performance the smallest of margins can mean winning or losing, and in today’s environment that could mean the difference between fame and fortune, or fading into obscurity.

 

F1 Eye Tracking

 

If we look at medical or clinical uses, being able to identify things like autism at an earlier stage (using non-verbal responses through measuring eye movements) allows parents and clinicians to adapt and plan a child’s education to minimise the impact on their development and lets the family be more prepared moving forward. Building up databases of typical and non-typical developing children from all walks of life both in and out of the lab allows milestones to be measured, new learning or rehabilitation techniques to be developed. Being able to extract information without the requirement for self reporting or verbal communication breaks down barriers that would otherwise mean that diagnosis may not be available for weeks, months or even years later otherwise. Using the latest techniques for training and both real and virtual presentation of scenarios means that we can now train healthcare professionals, surgeons and patients in situations that could be life threatening without the risk, and by understanding totally how they interact and engage gives us insights never before available.

 

Picture1

 

When looking at process management, health and safety or manufacture there are always people in a workplace that are ‘naturals’ at what they do, they have either adapted to their task very comfortable and excelled, or more likely through repetition and learning have become expert. Using eyetracking we can observe how these people operate, understand if and how they anticipate next steps, how they scan and search for elements or their situational awareness. Next we bring on the novice or the person to improve, observe them and compare them to our experts, guiding their interactions with a proven benchmark. An accident at work can be costly both in financial and possibly human terms, so use a simulator, VR environment or test area and monitor people’s actions and movements – and pre-empt possible bad situations. Does that fork lift driver check either side of the load often enough? How is that member of the QA team better at spotting defects in products – is their search strategy different? What makes that soldier better at finding ground disturbance in the field and locating IED’s? How can we be sure a mechanic checks every inch of an engine during a service and a vehicle is safe to use?

 

See how a mechanic checks an engine

 

Let’s think about consumer research – a mainstay of eyetracking and an ever growing market place. With the adoption of mobile devices on-screen real estate is smaller, we consume information quicker and we need to be more efficient at being noticed, getting our message across and of course helping the customer with their journey. A 1% increase in click-throughs, sign up or user experience could mean huge increases in a companies KPI’s but often selling in ideas and changes to a stakeholder can be challenging. Eyetracking provide a very visual way to demonstrate why customers aren’t (or indeed are!) doing what was expected on a website, image or menu system. Jumping into the retail space we are bombarded with products, signage, offers, POS, noise, colour and a whole lot more every time we walk through a shop entrance, or a mall, or a petrol forecourt – consumers self reporting their actions always has its limitations and this is even more evident in such a busy space as a retail outlet. Our eyes are digesting heaps of information, our brain is processing and discarding things that aren’t pertinent to the task and consumers simply can’t remember, never mind verbalise, all of this at the rate it is going. Unlock the subconscious by measuring the bodies leading input device – the visual system. Again small performance gains at the checkout in one store quickly multiply to large increases across a brand, retailer or globally. What distracts the shopper or draws their attention away from where we want them to look? What attracts them to our competitors? What elements do they use to navigate, make a decision or determine quality? Can people navigate around the virtual store before we invest in deploying the new layout?

 

Heat map in a virtual store

 

Think about your project, objective or study – is the interaction with the stimulus, product, environment or other people of interest? Do you want to know what and when they use visual information at any stage in the trial to inform the decision-making process? Do you want to understand why someone is better at a task than someone else? Do you want a very visible way of demonstrating a participants behaviour to a stakeholder? If the answer to any of these questions (or many more similar to these) is yes, then there is value in eyetracking for you.

Speak to us about methodologies for your study, the different types of equipment on hand and how we can help you get the insights you need.

 

Just What the Optician Ordered? November 25, 2015

Posted by Jon Ward in eye tracking, Glasses, Market Research, Shopper Research, Tobii.
Tags: , , , , ,
add a comment

It’s been a busy week in Sweden it seems as hot on the heels of the announcement of the amazing new 100hz speed option for the Tobii Glasses 2 system two new accessories have also been announced.

First up is the Precription Lens pack. The Prescription Lenses package contains corrective snap-on lenses for Pro Glasses 2 to facilitate studies involving subjects with either short or long sightedness. From mobile device testing and operator assessments, to sports research and wayfinding studies, these lenses will allow you to include a larger variety of subjects in your research.

The lenses range from -5 to +3 diopter in 0.5 diopter steps in order to provide support for a larger cross-section of the population with vision impairments. Separate lenses are provided for left and right eyes so you can accommodate differences between your subjects’ eyes.

Lens replacement is quick and easy with a magnetic frame that provides an easy-to-use, snap-on interface. At the same time, the solution is attached securely enough for test situations that involve a lot of natural movement, such as in sports research scenarios. Specific lens requirements and replacements are also available and the kit is supplied in a travel case.

 

Tobii Glasses 2 Prescription Lens Kit

 

Second is the new carry case and extended battery set. Combining a soft pouch with shoulder strap to carry the recording assistant and an external battery that neatly slots into its own holder on the bag the unit allows users to carry the recording unit in comfort and adding extra convenience for field based research with extended times between battery changes.

The case allows even quicker set-up and ease of use for the participant and researcher – and looks pretty good as well!

TG2 Pouch and Battery TG2 Pouch

Both accessories are available to order now – contact Acuity for more details.

Tobii Glasses 2 Real World Mapping – Saving Time in the Real World! October 29, 2015

Posted by Jon Ward in eye tracking, Glasses, Market Research, Marketing, Media, Shopper Research, Technology, Tobii, Updates, Usability & UX.
Tags: , ,
add a comment

The Tobii Glasses 2 have been a huge success, with the wide angled field of view, live wireless viewing and automatic slippage compensation thanks to the unique 4 camera binocular tracking more and more people are able to do some great fieldwork both unassisted and in a more traditional context. However as with all glasses based eye tracking platforms the analysis of the data is more time-consuming than with a screen based system as you need to code the user data onto reference images to create aggregated visual outputs and statistical metrics. This can be very time-consuming and depending on the type of interactions mapped (fixation or raw data) combined with the environment and task of the user (unboxing a product, retail purchases, driving and so on) can take anywhere from 10 to 20 times the duration of the recording to code – so 100 minutes of total recorded interactions could take upwards of 16 hours to code before you can begin the analysis.

Real World Mapping Example

Tobii Real World Mapping can help reduce this time considerably in many types of study by using advanced computer vision to automatically detect the gaze points based on the reference images you upload – meaning that 10x or 20x multiplier comes down to as little as 2-5x, and processing can be queued so the software can happily run along in the background freeing up valuable staff resources to focus on other tasks.Once the automatic process is completed you are presented with a chart plotting the automatically mapped points, alongside a confidence level in its accuracy and then any missed points (for example if there was a large amount of occlusion in the frame) or mapped points that need some adjustment can be manually corrected by a researcher.

Real World Mapping Example 2

This video has a brief overview of this exciting feature which will be part of next update of the Tobii Glasses Analyzer software.

Real World Mapping from Tobii

Of course not every single study will be able to take advantage of the new functionality, for example very dynamic content such as sports science studies have few or no fixed reference points to work with, objects that are largely occluded constantly or are at extreme distances will not be viable but for a large number of shopper studies, product interaction, mobile and tablet applications, advertising and navigation tasks users should see a significant time savings to using the tool – and by running a pilot (see tip 28 on Tim’s Acuity Intelligence blog for a reminder of the importance of this : http://www.acuity-intelligence.com/blog/eye-tracking-tips-26-30) you can quantify this benefit and also ensure that your reference images are correct and work well – more about that another time!

The Real World Mapping is ready to demonstrate to customers now – so get in touch and we will be happy to walk you through it, and show how it can help you process your glasses based eye tracking data quicker!

The New Tobii X2 Eye Tracker – The Smallest And Most Flexible Eye Tracker On The Market! February 11, 2013

Posted by Natasha French in Advertising, eye tracking, Market Research, Marketing, Media, Shopper Research, Technology, Tobii, Uncategorized, Updates, Usability & UX.
add a comment

Acuity are proud to present the new Tobii X2 eye tracker – a ground breaking development in delivering the smallest and most flexible eye tracker on the market!

Image

The Tobii X2-30 Eye Tracker (available in Compact Edition and Wide Edition) is a revolutionary small eye tracking system, powered by the latest generation in innovative eye technology from Tobii.

The Tobii X2 family comprises of eye tracking systems at 30 and 60 Hz. The X2 can easily be clipped on to a laptop, a PC monitor, or even a tablet for a compact and is our most portable system yet!

Research anywhereSmall footprint accommodates truly portable solutions and enables expansion of eye tracking from lab to real-life environments.

Supreme efficiency Ease of set up and operation paired with very robust participant tracking allow for cost efficient studies.

Trust your data – Unparalleled tracking accuracy within a revolutionary large head movement box ensures reliable and valid research results.

Choose between the Compact Edition and the Wide Edition – depending on your specific study context!

The Compact Edition is a smaller version of the eye tracker, measuring 184 mm (7.3’’) in length. You can use it as your portable lab or for studies that require a small eye tracker to track what participants see on:

  • Laptops and smaller PC monitors up to app. 22’’
  • Tablets and mobile phones (dedicated mobile device accessories will be available soon)
  • Small real-world interfaces

The Wide Edition is designed for studies that require larger gaze angles (up to 37°) and enables studies that involve larger stimuli, being able to track interfaces such as:

  •  PC monitors up to app. 27’’
  •  TV
  • Projections and simulators
  • Large real-world interfaces

Acuity are offering both rental and purchase options. As always for more information please contact the Acuity team at; sales@acuity-ets.com or (0)1189000795!

Why 2013 Is Going To Be Even Better For Our Clients…. February 5, 2013

Posted by Natasha French in eye tracking, Market Research, Shopper Research, Technology, Tobii, Uncategorized, Updates, Usability & UX.
add a comment

It’s been a busy couple of years with exciting developments here at Acuity. Key to this was the launch of Acuity Intelligence just over a year ago and the addition of Technical Director Dr Tim Holmes. I’ve had a few clients contact me, curious to find out the difference between the two companies: Acuity ETS and Acuity Intelligence (ETS and AI for short!). Reflecting on 2012 and what it held for the group, I thought the New Year might be a good time to clarify the difference between both and explain why 2013 is going to be even better for our clients.

Acuity ETS

Most of you reading this will already be aware of Acuity ETS. You’ll know our Directors and hopefully come to us when you have an eye tracking sale / rental requirement. Alongside this, ETS also offer training and support, giving you the tools and ‘know how’ to complete your own independent research effectively and use scientific technology to answer difficult commercial questions.

Now, if any of you know our Directors well, you’ll also be very aware that they are technology geeks (this enthusiasm is a must for anyone that joins the Acuity team!). With a (slightly obsessive) thirst for new tech, it wasn’t long before other tools that could supplement gaze data were also found by the guys. Wanting to offer these to clients too – the idea of Acuity Intelligence began!

Acuity Intelligence

Customer tracking / counting technology, facial and emotional recognition solutions, non-intrusive biometric sensors, browser based eye tracking to name but a few, the supplementary offering expanded quickly along with the AI team.

The end of 2011 welcomed a new company Director and Vision Scientist, Dr Tim Holmes. Actively involved in scientific research as a university lecturer, Tim was the catalyst for AI offering additional research and consultancy services based on real science. Services which include education about why technologies such as eye tracking and galvanic skin response are important measures of consumer attention and emotional engagement; the design of rigorous and innovative research projects that maximise the value of client investment by using the right combination of methodology and technologies to answer client questions; and innovative analysis and presentation of results that are scientifically accurate, replicable and focus on the meaning of the data, rather than just the numbers themselves. With Matlab and E-Prime consultancy offerings, AI also supports experiment development and data analysis in a number of Universities in the UK and beyond and is actively involved in cutting edge research at a number of institutions. In addition to all this, AI now also has product integrators and software developers in house who are feverishly beavering away on some truly unique off-the-shelf solutions as well as bespoke solutions tailored to the needs of specific customers!

Working together for you in 2013

Both sister companies offer something very different but both work together seamlessly on your behalf to provide you with a more complete research solution for 2013; all equipment rental, technical support and analysis or simply the technology and knowledge that will enable you to complete your own independent research well. We hope that by extending our offering to you, we can offer a more flexible way to make use of new technologies that add real scientific value to your research.

Your Chance to Win a Pair of Pivothead Glasses – Come See Us at Stand G77, The Ski and Snowboard Show 2012! October 29, 2012

Posted by Natasha French in Glasses, Market Research, Media, Shopper Research, Technology, Uncategorized.
Tags:
add a comment

Only 2 days until the Ski and Snowboard Show at Earls Court and I can’t wait! Scott, Tim, Alice, Nina and I are giving Acuity Intelligence a hand on stand G77 to showcase the AMAZING Pivothead Glasses!

With early starts and late finishes, it’s set to be a busy few days but with new show features, returning favourites, new experiences and spectacular stunts – oh, and did I mention a rock horror ‘Break The Ice’ party Thursday evening? It’s set to whet your appetite for your next adventure in the snow…

And what better way to capture all that snow adventure than by wearing the new Pivothead full HD 1080p hands free recording glasses available from www.acuity-intelligence.com These little gems are perfect for capturing all that daredevil snowboard and ski action and at 60 fps, you won’t miss a thing! Don’t believe me? Check it out at: http://goo.gl/1JHrZ Even better, come along and see us at Earls Court from 31st October to 4th November for your chance to win a set of your own Pivotheads!

For ticket information go to: www.skiandsnowboard.co.uk or for more information about the new Pivothead glasses contact the Acuity team at www.acuity-intelligence.com or (0)1189 000 820.