home..

Prac2 Data Annotation

Camera image annotation

In this practical, you will learn how to process and annotate the camera data you have been collecting over the past few hours.

0. Oxford wearable browser

Start by installing the Oxford wearable camera browser to your home folder. You can find more details about the camera browser on GitHub. In essence, it is a graphical user interface that allows researchers to annotate camera logger data. Note, this browser is currently in the process of being updated, so do not be surprised if there are dependency warnings in this legacy version.

Start by installing node.js on your iMac:

$ brew install node

Once node.js has been installed, clone the camera browser repository into your home directory:

$ cd ~
$ git clone https://github.com/OxWearables/oxford-wearable-camera-browser.git

Install the relevant dependencies, and ignore the dependency warnings.

$ cd oxford-wearable-camera-browser
$ npm install  

Finally, you can run the camera browser as follows:

$ npm start

It is import that you run npm start so that it sets up a folder called OxfordImageBrowser, which is where you will later add your camera data to. However, the browser does not work properly because we haven’t added data or schemas to the folder yet, which we will do shortly. You can exit the app by typing control+c in the terminal. Let us know if you have any issues here!

1. Data extraction

We now want to extract the raw images from your camera and get them into a format that can be used by the Oxford Image Browser.

We start by returning to the ~/practicals/scripts directory. We will again use autographer.py to download data from the camera.

$ cd ~/practicals/scripts
$ source .venv/bin/activate

Now plug in your camera. To download the photos, run:

$ python autographer.py --download True --destDir ~/OxfordImageBrowser/images/<your name>

To delete the photos on the device, run:

$ python autographer.py --delete True

Safely disconnect the device and return it to your tutor. If you open the folder (~/OxfordImageBrowser/images/<your_name>) where you extracted the images to, you can browse through the images and delete any that you wish to.

In order for the camera browser to work, we need to create thumbnail sized version of your photos. First install pillow, which is used for the image processing. Assuming you are still within the virtual environment in the ~/practicals/scripts folder, run:

$ pip install pillow
$ python create_thumbnails.py --root ~/OxfordImageBrowser/images/

We should now have the following directory structure in ~/OxfordImageBrowser/:

2. Annotating your data

Now we can begin to annotate our own data.

$ cd ~/oxford-wearable-camera-browser/
$ npm start

At this stage, you should see your name pop up on the left-most panel. If you click on it, you should see you data load to the browser area.

Now, select the annotation scheme annotation.csv. At this stage, it is worth explaining the annotation process in more detail.

All previous annotators of the CAPTURE-24 dataset had to go through extensive training - annotating at least 8 test subjects - before they can go on to annotate other data. We will not subject you to that experience! However, today you will get a taste of this by annotating your own data. To start with, please go through parts C & D of this supplementary document from this paper to understand how you should annotate the images.

The overall flow is described in great detail here where the essential elements are to:

  1. Divide up the images into different events/activities by clicking on the timeline above the images: I) To move event boundaries, click and drag the circles. II) For instance this split is where I think the participant has started locking up their bike.

  2. Look at consecutive images to determine the most suitable annotation for each image sub-sequence. You can also make use of the search bar if you are unsure where the annotations are located in the hierarchy.

  3. Pull the selected annotation onto the image(s). To change an activity annotation simply drag another annotation over it.

Once you have finished annotating your data, export the annotations to a .csv file (inside ~/OxfordImageBrowser/annotations/<your_name>/). This should happen automatically, but you can always do this manually by clicking on the download annotation button.

Exercise: Matching with accelerometer timings

So far this practical has streamlined data collection and processing. However, now we are going to get you to dive into some messy data processing. We are going to get you to match you accelerometer data with your camera annotations. In order to do this, you will need the .csv file of annotations, and you will need the .cwa file that you get from plugging in your accelermeter into your iMac. We will leave it up to you to come up with a suitable project structure to contain your data and your code (perhaps, take inspiration for the software engineering module). In order to process the accelerometer data, look at actipy, which is a python package for processing acclerometer data. It is up to you whether you want to do the data-processing in a notebook, or as a .py script. Matching up the annotations and accelerometer data will involve writing a function that matches each accelerometer data point, which has a time stamp, with one of the annotation labels, which is associated with a time range. However, this is not a one-to-one mapping since one annotation label might encompass multiple accelerometer readings.

At the end of this, you should have a numpy array of shape [T, 3] from the T accelerometer readings (each reading contains acceleration recordings in the X, Y and Z direction), a numpy array of shape [T] of the times that these readings were taken at and an array of shape [T] of the corresponding labels. For instance, if you have T=10 time readings, and have named the three arays accel, times and labels, then you get the follwing:

>>> print(accel.shape, times.shape, labels.shape)
(10, 3) (10,) (10,)

You can save these three files using:

>>> np.save("accel.npy", accel)
>>> np.save("times.npy", times)
>>> np.save("labels.npy", labels)

We recommend that you compress the file that you save your data to so that you can send it to the Virtual Machines (VMs) for the later practicals where you will be using the VMs and their GPUs for data processing. One can compress a folder (or file) using:

tar -zcvf my_data.tar.gz my_data

Then, to extract the file, type:

tar -xvf my_data.tar.gz

To send files between your local machine and a VM, one can use scp.

Exercise: Coming up with your own annotations

free_text.csv: In the previous exercises, you have been confined by the definition of the annotations to define ‘events’ / ‘activities’. Now come up with your own description or annotations and divide up the image timeline according to what feels most natural to you. In order to do this, put your event annotations / descriptions into the free_text.csv file, and drag these to annotate your events. You can edit this file within any tabular data editor, such as Microsoft Excel, or even just TextEdit. If you are confused about how to structure hierachical annoations, see the annotation.csv file. In essence, one uses semicolons ; to describe hierachical annotations.

Exercise: Identify the challenges with potentially noisy labels

One of the issues with annotations is that they may reflect the biases of the annotater. Typically, a well annotated data-set requires multiple annotaters annotating each data-point and some methodology for choosing the most likely label, such as choosing the majority label. In the case of annotating the CAPTURE-24 dataset, annotaters had to go through a training process that required them to obtain a kappa score of 0.8 relative to expert annotations on a subset of the data-set. Firstly, come up with some of the potential ways one can mitigate against biased labels, and also some of the implications of doing research using biased labels.

Optional: Cross-check your own annotations with others

One practical way one can try to quantify noise in the labels is by comparing multiple annotations of the same data. If you are willing to exchange data with someone else, and annotate each other’s data, the you can use this to compare your annotations with the other person’s. To run this script, you will first need to install four depedencies:

$ pip install argparse
$ pip install numpy
$ pip install pandas
$ pip install -U scikit-learn

Then generate the list of image file names in this format and store this as cdt-fileList.txt under the ~/oxford-wearable-camera-browser/training folder. Move your own annotation to the same place and rename it as cdt-ref.csv.

Now go a get another annotation file from your friend, you can run the comparison script via:

$ python kappaScoring.py /..path../annotation.csv

To generate the file list, you can use this Python script:

import os 
import glob

path2gen = '/Users/hangy/plots'
file_list = glob.glob(os.path.join(path2gen, '*.JPG'))

with open('file.txt', 'w') as f:
    for item in file_list:
        txt2write = item.split('/')[-1]
        print(txt2write)
        f.write("%s\n" % txt2write)

When you are done, let your tutor know what your kappa score is.

© 2023 Oxford Wearables Group