Copy the page URI to the clipboard
Tun, Thein Than; Bandara, Arosha K.; Price, Blaine A.; Yu, Yijun; Haley, Charles; Omoronyia, Inah and Nuseibeh, Bashar
(2012).
DOI: https://doi.org/10.1109/RE.2012.6345797
URL: http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arn...
Abstract
Privacy requirements for mobile applications offer a distinct set of challenges for requirements engineering. First, they are highly dynamic, changing over time and locations, and across the different roles of agents involved and the kinds of information that may be disclosed. Second, although some general privacy requirements can be elicited a priori, users often refine them at runtime as they interact with the system and its environment. Selectively disclosing information to appropriate agents is therefore a key privacy management challenge, requiring carefully formulated privacy requirements amenable to systematic reasoning. In this paper, we introduce privacy arguments as a means of analysing privacy requirements in general and selective disclosure requirements (that are both content- and context-sensitive) in particular. Privacy arguments allow individual users to express personal preferences, which are then used to reason about privacy for each user under different contexts. At runtime, these arguments provide a way to reason about requirements satisfaction and diagnosis. Our proposed approach is demonstrated and evaluated using the privacy requirements of BuddyTracker, a mobile application we developed as part of our overall research programme.