CSET 2013 : Workshop on Cybersecurity Experimentation and Test
Call For Papers
CSET invites submissions on the science of cyber security evaluation, as well as experimentation, measurement, metrics, data, and simulations as those subjects relate to computer and network security and privacy.
The science of cyber security is challenging for a number of reasons. For example, very little data is available for research use, and little is understood about what good data would look like if it were obtained. Experiments must recreate relevant, realistic features—including human behavior—in order to be meaningful, yet identifying those features and modeling them is hard. Repeatability and measurement accuracy are essential in any scientific experiment yet hard to achieve in practice. Cyber security experiments carry significant legal and ethical risks if not properly contained and controlled, yet often require some degree of interaction with the larger world in order to be useful.
Meeting these challenges requires transformational advances, including understanding the relationship between scientific method and cyber security evaluation, advancing capabilities of underlying experimental infrastructure, and improving data usability.
Topics of interest include but are not limited to:
Science of cyber security, e.g., experiences with and discussions of experimental methodologies
Measurement and metrics, e.g., what are useful or valid metrics, particularly when human behavior and perception (such as privacy) are considered? how do we know? how does measurement interact with (or interfere with) evaluation?
Ethics of cyber security research, e.g., experiences balancing stakeholder considerations, frameworks for evaluating the ethics of cyber security experiments
Alternative approaches to cyber security research, e.g., the application of methodologies from the social sciences (where observational experiments involving human behaviors are often used) to advance our understanding of cyber security phenomena
Data sets, e.g., what makes good data sets? how do we know? how do we compare data sets? how do we collect new ones or generate derived ones? how do they hold up over time? how well do red teaming or capture-the-flag exercises generate data sets?
Simulations and emulations, e.g., what makes good ones? how do they scale (up or down)?
Testbeds and experimental infrastructure, e.g., tools, usage techniques, support for experimentation in emerging security topics (cyber-physical systems, wireless, etc.)
Experiences with cyber security education, e.g., capture-the-flag exercises, novel experimentation techniques used in education, novel ways to teach hands-on cyber security
Because of the complex and open nature of the subject matter, CSET '13 is designed to be a workshop in the traditional sense. Presentations are expected to be interactive with the expectation that a substantial amount of this time may be given to questions and audience discussion. Some papers will be given their own time slot of about 45 minutes, while similarly themed papers may be grouped together for discussion. Papers and presentations should be conducive to discussion, and the audience is encouraged to participate. To ensure a productive workshop environment, attendance will be limited to 80 participants.
Position papers and research papers are welcome as submissions. Submissions that recount experiences (e.g., from experimentation or teaching) are especially desired; these submissions should focus on take-aways and lessons learned which might be helpful to other researchers conducting similar research. For all submissions, the program committee will give greater weight to papers that lend themselves to interactive discussion among attendees.
Research papers should have a clearly stated methodology including a hypothesis and experiments designed to prove or disprove this hypothesis.
Position papers, particularly those that are critiques of past work, should make certain to also include detailed proposed solutions.
Position and research submissions must be 6–8 pages long including tables, figures, and references. Text should be formatted in two columns on 8.5" x 11" paper using 10 point type on 12 point leading ("single-spaced"), with the text block being no more than 6.5" wide by 9" deep. Text outside the 6.5" x 9" block will be ignored.
All submissions must be anonymized. Blind reviewing of papers will be conducted by the program committee. Authors must make a good faith effort to completely anonymize their submissions. Submissions violating the detailed formatting and anonymization rules will not be considered for the workshop.
Submissions must be in PDF and must be submitted via the Web submission form, which will be available here soon.
All papers will be available online to registered attendees before the workshop. If your accepted paper should not be published prior to the event, please notify email@example.com. The papers will be available online to everyone beginning on the day of the workshop.
At least one author from every accepted paper must plan to attend the workshop and present the paper's findings.
Simultaneous submission of the same work to multiple venues, submission of previously published work, or plagiarism constitutes dishonesty or fraud. USENIX, like other scientific and technical conferences and journals, prohibits these practices and may take action against authors who have committed them. See the USENIX Conference Submissions Policy for details. Questions? Contact your program co-chairs, firstname.lastname@example.org, or the USENIX office, email@example.com.
Papers accompanied by nondisclosure agreement forms will not be considered. Accepted submissions will be treated as confidential prior to publication on the USENIX CSET '13 Web site; rejected submissions will be permanently treated as confidential.
|CCS 2013 Communications and Computer Security Journal|
|DATE 2013 Design Automation and Test in Europe|
|ACSAC 2013 29th Annual Computer Security Applications Conference|
|ITC 2013 International Test Conference|
|IJPHIM 2013 International Journal of Privacy and Health Information Management|
|DATE 2014 Design Automation and Test in Europe|