Over the last two decades the world has experienced a major grown in IT technology and cyber crime. The range of devices that have the ability to store and process electronic data is rapidly increasing, devices such as iPods, digital cameras, laptops, notebooks, tablets and smart phones are responsible for this. We are living in a era where mobile, instant access to data is crucial to our day to day lives. Unfortunately, these devices are not always used for sincere purposes, they are often found to be involved with a large number of crimes that occur each year.
With the above taken into account, it isn't hard to see why digital forensic analysts and experts are becoming inundated with devices. Numerous police forces' digital forensics units around the UK are suffering due to a backlog of electronic evidence (EE) collection and analysis. Wiltshire police were the highlight of a BBC Radio 5 Probe in 2015 which revealed that more than half of UK police forces were experiencing delays of at least 3 months for cases to be even allocated, "Five forces had devices which had not been examined after more than a year". (Gilbert, 2015)
With the above taken into account, it isn't hard to see why digital forensic analysts and experts are becoming inundated with devices. Numerous police forces' digital forensics units around the UK are suffering due to a backlog of electronic evidence (EE) collection and analysis. Wiltshire police were the highlight of a BBC Radio 5 Probe in 2015 which revealed that more than half of UK police forces were experiencing delays of at least 3 months for cases to be even allocated, "Five forces had devices which had not been examined after more than a year". (Gilbert, 2015)
To aid forensic investigators, there are various forensic software available out there in both commercial and open source markets. This is to increase the efficiency of investigations and also to increase the amount of devices that can be assessed than prior to using tools.
Unfortunately, these tools can vary in their performance, their cost and most importantly, their accuracy. Currently, there are no requirements for tools to be vetted or validated before used within a court case or by an expert witness. However, software validation standards have been around for almost as long as digital forensic investigations. Some of the many standards that exist are:
- ISO/IEC 17025 - General requirements for the competence of testing and calibration laboratories
- IEEE Standard for System and Software Verification and Validation. 2012
Whilst these standards exist for software to be validated and calibrated, they do not provide the entire methodology to testing a piece of software from start to end, no methodology can cover all grounds when it comes to validation of software.
Beckett (2007) expressed that bringing software to compliance with ISO 17025 is the best way to bring all other digital forensic tools (DFTs) to the same level of validity and reliability.
Research was performed into various software validation and evaluation methodologies that have been in use over the last couple of decades. Carl Erik Topp from NordTest (2014) created the TR 535 report which shows both a methodology and template report that implements the use of the ISO 17025 standards along with other recognized standards in software validation. This method was not chosen however as it was much more low-level than required for this assessment, going into depth regarding scope and low-level computing validation.
Daniel Ayers (2009) produced an article in proceedings of the ninth annual DFRWS conference regarding the current state of software validation of DFTs. He proposed several new requirements for the "second generation of DFTs" along with several evaluation metrics that should be considered during an assessment of a DFT.
Requirements from Ayers (2009):
- Parallel processing
- Data storage and I/O bandwidth
- Accuracy and reliability
- Auditability
- Repeatability
- Data abstraction
Evaluation metrics:
"Absolute speed – the elapsed (wall clock) time required to complete analysis.
Relative speed – the average rate at which the tool is able to process evidence compared with the rate at which data can be read from the original evidential media.
Accuracy – the proportion of analysis results that are correct.
Completeness – the proportion of forensic artefacts present in the evidence that are identified and reported by the tool.
Reliability – the proportion of tests where the tool executes successfully, does not crash or hang, and provides output in the documented format. (Note that this metric is not concerned with the accuracy of results.)
Auditability – the proportion of results which are fully auditable back to original evidence data, including documenting all computations performed to derive results and all assumptions and other inputs (such as configuration information set by the analyst) that are capable of influencing results.
Repeatability – the proportion of tests where it can be established, for example using detailed logs generated by the tool, that the process employed for analysis of an evidence item was exactly the process specified." (Ayers, 2009)
With the above taken in to account, for a valid attempt at evaluation software, a process (or methodology) must be followed to ensure thoroughness and completeness of validation.
Test Data Samples
There are multiple resources available for forensic reference data sets, such as the Computer Forensic Reference Data Sets (CFReDS) project by NIST. (2016) This project provides anonymous test data with detailed expected results. These generally appear to be for memory forensics and data imaging / recovery tools. No browser profiles were found during research of the various data sets that are provided by the CFReDS project nor external links from this project.
Whilst efforts were applied to finding browser profiles that matched Firefox, SeaMonkey and IceWeasel, none were found.
Due to this, example user data profiles will be created with defined expected results, for example,
when testing the ability of listing the users cache from Dumpzilla, several pages will be loaded and the web developer tab will be used to capture all files that are being cached by the browser. This can then be cross-referenced with the results of Dumpzilla.
A total of 5 test profiles will be created.
- 3x Firefox profiles with various expected results per assessed function
- 1x IceWeasal profile with at least one expected result per assessed function.
- 1x SeaMonkey profile with at least one expected result per assessed function.
Testing Process:
The below process will be followed when evaluation each function of the Dumpzilla DFT ( each function defined in another post).
- Virtual Machine is reverted to a previous snapshot before any testing had commenced.
- The test data is created via loading up the relevant browser and performing operations to input data into analyzed datastores.
- All background applications are closed down and the system is rebooted to clear system cache and volatile memory space.
- The Dumpzilla script is then used, pointed at the current test profile. With one function used per run.
- Time of process is captured (Absolute speed)
- Results are logged for comparison (Accuracy, Completeness)
- Any errors are logged for reliability
- The step is repeated for a total of 3 times for a larger result set.
- Results are collated per browser and function.
- Results are compared for accuracy of the expected results.
- Any anomalous results are further investigated for root cause.
This blog will continue to be updated over the course of the next month showing the various stages of the testing process and creation of test data. All results are expected to be public by the 27th of November 2016.
Bibliography
Gilbert, D. (2015) Police force delay in forensic examination revealed by BBC probe. heraldscotland [Online], 9 November. Available from: <http://www.heraldscotland.com/news/> [Accessed 01/11/2016].
International Organization for Standardization (2005) ISO/IEC17025 General requirements for the competence of testing and calibration laboratories Second Edition [Online]. Switzerland: ISO. Available from: <http://www.uobaghdad.edu.iq/uploads/pics13/q1684/iso17025_eng.pdf> [Accessed 01/11/2016]
J. Beckett and J. Slay. (2007) Digital forensics: Validation and verification in a dynamic work environment: proceedings of the 40th Hawaii International Conference on System Sciences, January 3-6, 2007, Hawaii: IEEE Computer Society
National Institute of Standards and Technology. (2015) The CFReDS Project [Online]. Available from: <http://www.cfreds.nist.gov/> [Accessed 01/11/2016]
National Institute of Standards and Technology. (2015) The CFReDS Project [Online]. Available from: <http://www.cfreds.nist.gov/> [Accessed 01/11/2016]
NordTest. (2014) History - Nordtest.info [Online]. Available from: <http://www.nordtest.info/index.php/nordtest/history.html> [Accessed 01/11/2016]
Ayers, Daniel. (2009) Digital Investigation: A second generation computer forensic analysis system: proceedings of the Ninth Annual DFRWS Conference, September, 2007 Montreal: Elsevier Ltd
Ayers, Daniel. (2009) Digital Investigation: A second generation computer forensic analysis system: proceedings of the Ninth Annual DFRWS Conference, September, 2007 Montreal: Elsevier Ltd






