Using FSR based muscule activity monitoring to ... - Semantic Scholar

4 downloads 45704 Views 176KB Size Report
3. computer (gestures 9-12) It starts with the opening of a laptop. After that some random text is entered via the keyboard. Then some random mouse movements.
Using FSR based muscule activity monitoring to recognize manipulative arm gestures Georg Ogris, Matthias Kreil, Paul Lukowicz Embedded Systems Lab (ESL), University of Passau {georg.ogris,matthias.kreil,paul.lukowicz}@uni-passau.de

Abstract We present an experiment that investigates the usefulness of muscle monitoring information from arm mounted force sensitive resistors (FSR) for activity recognition. The paper is motivated by previous work that has demonstrated the feasibility of using FSRs for muscle activity monitoring (on leg muscles) and presented some initial signals related to distinct arm activities. We systematically investigate the performance of an FSR system on 16 different manipulative gestures and 2 subjects. The aim is to test the limits of the system, compare them to established sensing modalities (3D acceleration and gyro), and establish the value of combining FSR with other sensing modalities. We also present a hardware setup that addresses key problems that were identified in previous work: large variations in the attachment force and sensor placement accuracy issues. For all classifiers the overall accuracy of the FSR system is in the middle between the accelerometer (between 5% and 10% better) and the gyro (between 2% and 11% worse). Adding FSRs to another sensor improves the accuracy by 1% to 29%.

1

Introduction

Arm and hand actions are a key component of user activity. As a consequence much research in wearable context recognition has gone into tracking and recognizing such action. The vast majority of this research is based on motion sensors placed at different arm and hand locations (e.g. [7, 2]). This has both advantages and disadvantages. On the positive side, motion sensors, in particular accelerometers, are cheap, small and low power. At the same time a significant amount of activity information is contained in motion patterns. On the negative side, motion sensors tend to contain a mixture of motion and orientation information that is difficult to separate with simple sensor setups. In addition they provide no information about the palm and finger activity. Alternative approaches that address the above disadvantages include video tracking of the arms [5] and sensor

gloves. Our group has also investigated the use of ultrasonic beacons to track hand positions [6]. General Idea This paper investigates an additional source of information about arms and hand actions: the analysis of arm muscle activity with force sensitive resistors (FSRs). FSRs are thin piezoelectric plates (see figure 2) that change their electrical resistance as mechanical force is applied to their surface. They are cheap and can be easily integrated in garments. It has also been shown ([4]) that such mechanical pressure sensors can be implemented directly into textiles. The idea behind our work is to use such sensors in an elastic sleeve mounted on the forearm. It is motivated by the fact that palm and finger motions are driven by muscles in the forearm. As those muscles contract they change their shape which in turn results in mechanical pressure being applied to the sensors in the elastic sleeve. Paper Contributions In previous work [3] we have demonstrated the general feasibility of using FSRs to monitor leg muscle activity. In [1], we have also shown that different arm actions such as holding a heavy object or making a fist produce distinct FSR signals. This paper goes beyond such simple signal examples to investigate using FSRs for actual recognition of a set of non trivial manipulative gestures. The aim is not to demonstrate the ability to reliably recognize a specific, application relevant activity set. Instead we want to lay the groundwork for other researchers to set up a FSR based activity recognition system. To this end we present the following specific contributions: 1. We systematically investigate the performance of an FSR system with different classifiers on a set of 320 manipulative gestures from 16 different classes performed by 2 subjects. The classes have been chosen to test the limits of the system rather than be recognizable with high accuracy. Thus, the set contains some very subtle gestures. 2. We compare the FSR performance to established sensing modalities (3D acceleration and gyro). To cap-

Figure 1. Contestant wearing the FSR sensor system, a

Figure 2. Schematic of current to voltage converter, small

MTx glove, and a mobile computer, that serves as a power source.

left: addon boards for the Tmote Sky, small right: FSRs

ture hand rather than only arm motions (as do the FSR through muscle monitoring) the additional sensors are mounted on the back of the hand rather than on the wrist. 1 3. We investigate the benefit of combining FSRs with additional sensing modalities by testing different combinations of the three sensor types (acc, gyro, FSR). 4. We describe in detail a hardware setup that addresses key problems that were identified in previous work: large variations in the attachment force, and sensor placement accuracy issues.

2

The Sensor System

Sensor Placement and Attachment In [3] our group has described the effect of sensor displacement on muscle activity monitoring. It was shown that sensor displacement of just 1 cm can lead to false or no signal. However we have also demonstrated that this problem can be overcome by covering a larger area with the FSRs. This can either be accomplished by a matrix of sensors around the point of interest (as discussed in [3]) or by using large FSRs. In this paper we combine both approaches. Both the lower part of the forearm (right behind the wrist) and the upper part of the forearm (right under the ellbow) are covered with a ring of 4 46 x 46 mm FSRs. The attachment of the sensors is another crucial question. FSRs require a moderate counterforce. On the other hand, the system should be wearable, i.e., easy to put on and off and not too tight. As a consequence we opted for a three layer design: a thin inner layer (a thin stocking) on which the FSRs - the second layer - are fixed, and a third outer layer that is tight but stretchable. 1 Wrist mounting is actually more common in application since attaching devices to the hand is mostly considered burdensome (requires at least a glove). However we know from previous work that wrist mounted motion sensors are not good in recognizing gesture primarily defined by hand motions, so that hand mounted sensors offer a more challenging comparison.

We considered an ordinary bicycler’s sleeve to be the right choice. That way we end up with a sleeve that can be put on and off easily. Sensor System Even with the above attachment concept there are significant, user dependent variations in the FSR counter forces. When measuring the resistance of the FSR by means of a voltage divider as done in previous work, see e.g. [1, 3], the initial counter force must be adapted for each user due to the nonlinear relationship of resistance and output voltage. To overcome this problem the new system is based on current measurement instead of voltage measurement. That requires a current/voltage converter (Figure 1) but results in a linear relationship between resistance and output voltage with a much better dynamic range. Thus we get usable FSR signals despite the variations in the counter force. A simple standard motion (e.g. bending the arm) performed after the sensor have been put on can be used to calibrate the system to a specific user. The overall sensor system (Figure 2) comprises a tmote sky from moteiv with addon boards mainly featuring an ADG708 to multiplex the 8 FSR channels and the current/voltage converter. Future work An important next step will be to investigate in more detail and possibly automate the calibration of the system to different user. An auto calibration procedure could also compensate changes in signal baseline caused by slipping sensors.

3

The Experiment

The experiment design was not driven by a specific realistic application. Instead we aimed to have an activity that contained a reasonable variety of realistic gestures that would test the limits of what can be achieved with FSR muscle activity monitoring. In particular we wanted to have a mixture of bold and subtle gestures, gestures performed with different force and a different amount of ’gripping’.

1 2 3 4 5 6 7 8

open the pen write on whiteboard close the pen erase scroll screen down point on screen scroll screen up press buttons on projector remote control

9 10 11 12 13 14 15 16

classifier HMM C4.5 kNN

open notebook type a few words use the mouse close the notebook open bootle pour water in a glas close bottle drink

Table 1. Set of gestures At the same time we wanted to have a situation where the gestures are performed in a natural way as part of an every day activity rather than make the subjects perform a set of artificial gestures. Furthermore we wanted a scenario where the technical and organizational effort involved in the experiment is kept to a minimum needed to collect a solid, representative data set. Experimental Scenario As a consequence of the above considerations we have opted for a scenario of giving a talk at a seminar. Table 1 lists the gestures performed. They can be organized in 4 groups of actions consisting of 4 different subroutines. 1. whiteboard (gestures 1-4) First the test person opens a pen, then he starts to write some random words on the blackboard. After that he closes the pen, which is similar to the opening movement. Finally the words are cleared with an eraser, which is usually done in a periodic circulate movement. 2. beamer screen (gestures 5-8) This gesture set contains first the pushing down of the screen for, e.g., a beamer. Then the test person points on the screen with a finger to show some things. After that the screen is lifted up again, which is similar to the first movement. The last action is using the remote control to switch off the beamer by pressing keys. 3. computer (gestures 9-12) It starts with the opening of a laptop. After that some random text is entered via the keyboard. Then some random mouse movements are done. Next the laptop is closed again, which is a similar movement to the opening of the laptop. 4. drinking (gestures 13 - 16) A bottle of water is first opened by unscrewing the lid. Water is then poured into a glass and the bottle is closed again. Finally the test person takes a sip of water from the glass. Experimental Setup We mounted our FSR sleeve on the right forearm. Furthermore the subjects wore a glove with a MTx sensor from Xsens. The MTx sensor is mounted on the back of the hand rather then on the wrist to prevent its housing from influencing the FSRs. Note that this provides significantly more information on palm actions then

acc 83 76 81

gyr 72 57 65

fsr 73 62 76

acc+gyr 91 82 90

fsr+acc 84 79 86

fsr+gyr 81 68 84

Table 2. Classification results in % the more typical (and much less obtrusive) wrist based setup and to a degree neutralizes the inherent advantage of FSR. The FSRs were sampled at a rate of 20 Hz, the MTx at 50 Hz. In total we recorded 2 subjects, with 10 datasets each and one instance of every activity listed above per dataset (total of 320 gestures). The experiment environment was a meeting room at the university with a whiteboard, a beamer screen and a meeting table. Recognition Method For comparison, three classifiers have been tested. The tree based C4.5 classifier and the instance based k-Nearest-Neighbor (kNN) are used in a sliding window approach: In a time window of fixed size, a set of features is computed using the raw sensor data. Then the sliding window is moved by an offset which determines the overlap with the last window. We used mean and variance as features, with windowsize 30 and stepsize 15. After that a majority decision was applied to the raw classification results. This yields a filtered decision for the particular gesture and constitutes the final result of the frame based classification. We used the YALE implementation of these classifiers. In addition to this frame based approach a Hidden Markov Models (HMM) based classifier is tested as well. For each manipulative gesture in our experiment a separate HMM is trained. During testing a single gesture is aligned with the most likely model. We used the HMM implementation in the Bayes Net Toolbox for Matlab for our experiments.

4

Results and Discussion

Results Due to the small data set we evaluated the data in a cross validation scheme. The classification results for the three classifiers are summarized in Table 2, a per class recognition rate for different sensor modalities for the kNN classification is given in Table 3, for the HMM classifier in Table 4. The following are the main observations: a. The recognition is far from perfect for all combinations of sensors and all classifiers. This was to be expected, as the gesture set was chosen to test the limits of the recognition rather than to be fully recognizable. Typically most errors occur for do/undo gesture pairs. Such pairs of classes are: C1,3 , C5,7 , C9,12 , C13,15 .

sensor

P

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

acc gyr fsr

81 65 76

80 58 54

99 100 93

58 10 55

100 99 100

73 80 85

75 73 90

79 64 74

100 93 100

61 11 45

99 46 87

93 90 88

63 31 34

69 51 59

93 94 82

55 46 80

100 99 98

acc+gyr fsr+acc fsr+gyr

90 86 84

91 71 66

100 100 96

63 61 65

100 100 100

94 85 89

92 100 95

81 78 78

100 100 100

80 77 74

97 96 92

98 97 91

83 62 48

89 76 80

100 99 92

82 87 84

100 100 98

0.7 0.6

0.4 0.3

V

ADC

[V]

0.5

0.2 0.1

Table 3. Recognition results per class for the kNN classi-

0

fication for different sensor modalities.

0.6

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

acc gyr fsr

83 72 73

89 43 63

95 93 79

60 35 51

98 96 100

73 72 74

99 86 81

79 76 76

100 79 99

58 32 53

99 87 51

98 94 70

67 56 68

64 68 74

90 88 76

66 59 63

95 87 88

acc+gyr fsr+acc fsr+gyr

91 84 81

87 74 56

96 98 88

69 55 58

100 100 100

92 83 81

99 96 91

88 77 81

100 100 100

74 74 66

98 76 73

100 83 86

85 74 73

95 82 76

100 93 93

74 80 77

97 94 91

Table 4. Recognition results per class for the HMM classification for different sensor modalities.

b. For all classifiers the overall accuracy of the FSR system is in the middle between the accelerometer (which is between 5% and 10% better) and the gyro (which is between 2% and 11% worse). This is also not surprising, since the accelerometers are mounted on the hand rather than the wrist. Thus, just like FSRs, they provide information not just on motions, but also on grasping which causes vibration on the back of the hand. At the same time the FSRs have less clear signals due to placement issues and what is more lack some motion information. On the other hand gyros lack the grasping information (reaction to the vibrations is minimal). c. Adding FSRs to other sensors always leads to an improvement (between a minimal 1% for HMM and acceleration and 19% for kNN and gyro). This is clear for gyros. For accelerometers it indicates that there is indeed some information that the FSRs have and an accelerometer - even when hand mounted - does not have. This is further confirmed by the fact that even in a single sensor case there are gestures for which the FSR is better than accelerometers.

Lessons Learned The main lesson of this work is that FSR based muscle monitoring is indeed useful for the recognition of activities involving hand actions. While inferior to accelerometers mounted on the hand for the overall accuracy on the 16 gestures, FSRs perform well for many individual gestures. On a few they are even better then the accelerometer, which confirms that there is some grasping related information that even hand mounted motion sensors can not detect. Note, that in our experiment FSRs rely on much less obtrusive arm mounting which makes them preferable for many applications.

0.4 0.3

V

P

ADC

[V]

0.5

sensor

0.2 0.1 0

0

2.5

5

7.5

0

1

2

0

1

2

3

time [s]

Figure 3. Signal samples for the upper 4 FSR channels, left column: 2 samples from class drink, middle: open notebook, right: close notebook.

Limitations Although the gesture set in the experiment has been chosen to be diverse, it cannot be claimed to be representative in any systematic way. In addition, even for the same gesture set the performance for a given sensing modality depends on fine tuning of features, window sizes, etc. which were not explored systematically in this work. Thus the lessons and conclusions discussed above must be taken as indicative rather than proven beyond doubt.

References [1] O. Amft, H. Junker, P. Lukowicz, G. Tr¨oster, and C. Schuster. Sensing muscle activities with body-worn sensors. In BSN 2006: Proceedings of the International Workshop on Wearable andImplantable Body Sensor Networks, April 2006. [2] L. Bao and S. Intille. Activity recognition from userannotated acceleration data. In Proc Pervasive Computing, 2004. [3] P. Lukowicz, F. Hanser, C. Szubski, and W. Schobersberger. Detecting and interpreting muscle activity with wearable force sensors. In Pervasive 2006, pages 101–116, May 2006. [4] J. Meyer, P. Lukowicz, and G. Tr¨oster. Textile pressure sensor for muscle activity and motion detection. In ISWC 2006: Proceedings of the 10th IEEE International Symposium on Wearable Computers, October 2006. [5] T. Starner, B. Schiele, and A. Pentland. Visual contextual awareness in wearable computing. In IEEE Intl. Symp. on Wearable Computers, pages 50–57, Pittsburgh, PA, 1998. [6] T. Stiefmeier, G. Ogris, H. Junker, P. Lukowicz, and G. Tr¨oster. Combining motion sensors and ultrasonic hands tracking for continuous activity recognition in a maintenance scenario. In 10th IEEE International Symposium on Wearable Computers (ISWC), October 11-14 2006. [7] J. Ward, P. Lukowicz, G. Tr¨oster, and T. Starner. Activity recognition of assembly tasks using body-worn microphones and accelerometers. IEEE Trans. Pattern Analysis and Machine Intelligence, 28(10):1553–1567, October 2006.

Suggest Documents