Home

Benchmarking of filtering software and services

image

Contents

1. Effectiveness Over Blocking Localisation Table 3 FCM Model for Filtering tools Benchmarking The defined Tree for QUALITY is an adaptation of the 1509126 specification to the specific project s needs Each of the defined criteria is mapped to a number of metrics Each of the single metric value is normalised in order to convert its value in the range 0 1 In the standard FCM model there are three different filters that can be associated with each metric linear exponential and gaussian Each filter can be adjusted to perform the required conversion by properly setting the values for the parameters and b see the following picture Further filters and metric weightings may be added during the test development phase 1 1 1 0 5 0 5 0 5 0 0 0 b a a a b ab a ath Linear filter Exponential filter Gaussian filter f a 1 f b 0 f a 0 99 f a b 0 5 f a 1 f a b 0 5 6 3 Software Quality Factors and Criteria In the following section we define the metrics that have been assigned for each quality subcharacteristic Some of these criteria have been adopted directly from the ISO 9126 Benchmarking ot tiltering sottware and services Evaluation Criteria recommendations others have been introduced to meet the needs of the benchmark evaluation process For each set of quality criteria a set of measurement metrics have been defined these are discussed in section 7 6 3 1 F1 Func
2. Benchmarking of filtering software and services An Analysis Framework Definition of the Evaluation Criteria Version Issue 1 Draft 2 Benchmarking ot tiltering sottware and services Evaluation Criteria Title Benchmarking of filtering software and services Definition of the Evaluation Criteria Abstract This document provides the definition of the evaluation criteria for a benchmarking process for filtering tools and services Issue Issue 1 Draft 2 Date February 16 2001 Authors Tom Jackson Marco Riva Francesco Puglisi Joint Research Centre TXT e Solutions Distribution Unlimited The role of the Joint Research Centre of the EC is to provide scientific support to the EU policy making process by acting as a reference centre of science and technology for the EU This report has been prepared by the Joint Research Centre in the frame of its institutional support programme to the EC DG Information Society The opinions and views expressed in this report do not represent the official opinions and policies of the European Commission We invite readers of this report to send comments or suggestions to Tom Jackson TP210 ISIS Joint Research Centre ISPRA VA 21020 ITALY Tel 39 0332 786250 Benchmarking ot tiltering sottware and services Evaluation Criteria 1 2 3 4 8 INTRODUC HON EO 4 1 1 STUDY CONT
3. 6 3 3 3 F3C3 Friendliness Software attributes affecting the effort made by the user in order to access to all the software features e g number of clicks needed to perform an operation or easy and intuitive access to all the features from a menu 6 3 3 4 F3C4 Operability Software attributes affecting the effort made by the user in operating and controlling the software 6 3 4 F4 Effectiveness This category is has been introduced into the ISO 9126 model to specifically handle the analysis related to the filtering performance of the tools There are three quality criteria 6 3 4 1 Blocking Performance Software attributes relating to the performance of the tool in correctly blocking harmful content 6 3 4 2 Over blocking Performance Software attributes of the tool that lead to incorrect blocking of internet material that has no harmful content 6 3 4 3 Localisation Software attributes that relate to the ability of the tool to adapt to different cultural and linguistic operating environments 7 Definition of the Measurement Metrics In the following section we define a list of measurement metrics associated with each quality subcharacteristic This is an initial proposal It is expected that the metrics will be refined and where necessary extended during the course of development of the evaluation process and through open peer review ar Benchmarking ot tiltering sottware and services Evaluation Criteria 7 1
4. Comprehensive software evaluation is a time consuming and expensive process To date most filtering software trials have deployed relatively limited testing of the blocking performance e g the tools are typically tested on less than 100 URL s because the evaluations have used a manual test process It is expected that the mechanism for evaluating both blocking and over blocking performance see section 4 1 for a definition of these terms can be automated via web browser scripts deploying standard programming methods such as JAVA or XML These mechanisms will facilitate much broader test coverage for filtering performance 4 The Benchmark Process In the following discussion the structure of the benchmark process is explained The benchmarking process has three principal components 1 A set of evaluation criteria defining the measurements to be applied during the test process 2 A test method that defines how the tests should be performed 3 data processing method that defines how the results of the tests should be processed 4 1 Evaluation Criteria The evaluation criteria define the measurement framework for the benchmarking process Following a review of the literature on filtering testing and user requirements it has been decided that the following assessment areas must be addressed by the benchmark e Blocking effectiveness e Over blocking sensitivity e Security integrity e Operational integrity e Configurability e C
5. Functionality Usefulness Metrics URL blocking Protocol blocking TCP IP Word blocking ICQ chat blocking newsgroup blocking Email program blocking Outlook Express Opera Email attachments blocking send receive Email blocking download blocking Application blocking File blocking Control Panel access blocking Rating system blocking PICS Access time blocking Flexibility URL Blocking Flexibility Protocol Blocking Flexibility Keyword Blocking Flexibility Email Blocking Flexibility Download Blocking Flexibility Rating system Blocking Flexibility Access time Blocking Flexibility Complete URL definition Partial URL definition regular expression or other method Ability to redirect to a specific port Complete word definition Partial word definition regular expression or other method Use of predefined words Ability to define blocking for specific accounts PDF Napster ZIP DOC Multimedia Applications e g Naptster Real Audio MP3 MPEG Quicktime SafeSurf or RSACi Activation Scheduling of allowed forbidden time definition of maximum connection time Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No 47 Benchmarking ot tiltering sottware and services Evaluation Criteria 7
6. and the test results are largely derived from ad hoc test methods A study that is herd relevant to this benchmarking exercise is the work undertaken by the project a Preparatory Action for the Safer Internet Action Plan This study evaluated the effectiveness of commercial software filtering tools and their suitability for European users The study which was largely based on pa an European end user survey highlighted many of the current limitations of filtering tools and services from the perspective of the users These included the following major issues e Difficult to install and configure for non expert PC users e Filtering performance is erratic harmful content is often not filtered and acceptable sites are inexplicably blocked e diverse cultural and linguistic aspects of EU are not catered for by the tools for example offensive text is only recognised and filtered if written in the English language 3 The Benchmark Proposal In order to stimulate the development of improved filtering tools and services and in order to improve awareness of the capabilities and limitations of filtering software it has been proposed that a benchmarking process should be developed The benchmark process will achieve these objectives though a number of distinct mechanisms e Performance Goals By developing an independent and standardised test process that can rigorously evaluate the performance of filtering tools in all key
7. e g performance behaviour of software filtering tools The measurement metrics must be developed such that quantitative measurements can be made both for functional and non functional properties This may present difficulties in some assessment areas most notably those associated with usability Usability is a highly subjective software property dependent upon diverse criteria such as user experience user background operating environment and personal preference Consequently usability issues will not focus on the operational and user interface aspects of the software but rather on the more objective functional properties that allow the user to interact with the software tool and that determine the user s perception of a tool Here we include issues such as traceability transparency and cost effectiveness 44 Benchmarking ot tiltering sottware and services Evaluation Criteria 6 Filtering Software Quality Assessment The benchmark filtering process should become a standardised test method for Europe This will ensure that European end users have access to consistent evaluation information relating to the performance of filtering tools To facilitate the standardisation of the benchmark process it has been decided that the test should fit within the framework of known and established software evaluation standards To this end the study has adopted the ISO 9126 standard for software quality assessment 6 1 1 ISO 9126 Software
8. 17 7 1 1 Useumess Metis 17 7 1 2 Flexibility on e 17 7 1 3 Interoperability 5 i e eee eree eed orisirisii 18 7 2 WIS ABIETDY secs 18 7 2 1 Underst andability 18 7 2 2 18 7 2 3 REQuireMentS cccccssssscccceceesesssccececeeesessaeeeecceesesssaneeececeeesesaeeeeeeseessseaaeeeeeeeess 19 7 2 4 OPEV GDL A AA ES E E E A 19 7 3 ESNE E EE NEE E EASE O NEE E E EE EE EEA 20 7 3 1 EEE E ET 20 7 3 2 coi bets Soa TR 20 7 3 3 21 7 4 55 0 Beats sd eae 21 CONCLUSIONS neie osoasa ineeie osora saoao eseas SEE ENEE ASCE S ESOS SS snstcocsesecesessetdseudssecesssetdccossrsessesees 21 Benchmarking ot tiltering sottware and services Evaluation Criteria 1 Introduction This document is an interim deliverable of the study Benchmarking of Filtering software and services An Analysis Framework It sets out a first draft of a set of evaluation criteria to be used in a benchmarking process for filtering software and services The objectives of the benchmarking process are set out in the study Concepts Document and a brief
9. Good lists In principle the filtering performance in blocking harmful content should be perfect when using approved lists dependent upon the subjective criteria used by the software suppliers to define acceptable content However many users have stated that deploying only approved URL lists also known as a walled garden approach creates a browsing experience that it too restrictive for all but the youngest of children Consequently the test process will be conducted on the assumption that the tools should be configured for the trials so that they deploy only blocked URL lists not approved lists Tools will also only be tested with the lists supplied by the software providers Additional user defined lists or third party sourced lists will not be used within the benchmark trials Which Keyword List The filtering performance of the tools is also highly constrained by the keyword list that is deployed by the text based matching algorithms The keyword list is the list of prohibited harmful words that a tool attempts to identify and block when examining a web page for content Access to a web page is typically denied if one of the prohibited words on the list is identified on a page Some tools come with preconfigured lists others are user definable The test process will be conducted on the assumption that all of the tools will be configured with the same set of prohibited keywords Which Domain list The argument identified for
10. Product Evaluation The objective of the standard is to provide a framework for the evaluation of software quality ISO IEC 9126 does not provide detailed quality requirements for specific software products but it defines a quality model which is intended to be applicable to all types of software applications It defines six product quality characteristics which are described in figure 2 Figure 2 The Quality Factors defined by ISO IEC 9126 Are the required functions available in the software Functionalit ility How easy is it to transfer the software to another environment Maint How easy is it to modify the software Efficiency How reliable is the software Is the software easy to use How efficient is the software 7 ISO IEC 9216 Information Technology Software Product Evaluation an Benchmarking ot tiltering sottware and services Evaluation Criteria ISO 9126 stipulates that the quality factors are divided into a number of quality subcharacteristics or criteria These are described in table 2 Quality Factor Quality Subcharacteristics Suitability Accurateness Functionality Interoperability Compliance Security Maturity Reliability Fault Tolerance Recoverability Understandability Learnability Operability Usability Time Behaviour Resource behaviour Analysability Changeability Stability Testability Adaptability Installability Co
11. superkids com aweb pages reviews kidsafe 1 sw_sum1 shtml Access Denied The Impact of Filtering Software on the Lesbian and Gay Community GLAAD Dec 1997 Smut Filter Blocks All But Smut Wired http www wired com news technology 0 1282 36923 00 html Free speech advocate raises ire of filtering firms CNET News com http news cnet com news 0 1005 200 1567022 html Filtering the Internet A Best Practices Model J M Balkin B Noveck K Roosevolt Yale Law School Yale University New Haven USA http stiftung bertelsmann de internetcontent english download Filtering doc Benchmarking ot tiltering sottware and services Evaluation Criteria weaknesses inherent in the existing software tools and services Aside from political and cultural issues these criticisms are focused principally on the technical limitations of URL and keyword filtering These limitations can be largely attributed to the problems of determining the context of the data being filtered The tools have been shown to exhibit problems of adequately blocking harmful content and of also being prone to incorrectly blocking acceptable material Technical problems have also been identified with content labelling approaches These include questions as to who should do the labelling and how to ensure consistency of labelling across cultural boundaries These evaluation exercises have also highlighted the fact that there are no standard testing approaches for evaluating filtering tools
12. synopsis is provided below 1 1 Study Context The invention of the World Wide Web and of the graphics enabled browser were the catalysts for a vast explosion of the availability of Internet content and of the number of Internet users world wide who now include an increasing proportion of families with children and schools which provide access for pupils However with the explosion of available information has come the problem of undesirable or illegal data content which is inappropriate for access by minors One of the most widely publicised areas of undesirable material is the availability of so called adult content that is to say commercial pornography of various degrees most of which is legal for adult consumption in the country of production However the Internet also serves as a distribution mechanism for other forms of undesirable or illegal content such as racism violence and paedophilia Surveys have shown that consumers parents schools libraries etc view the unrestricted access to this material with great concern Access control tools providing content filtering offer one approach to meeting concerns of consumers These approaches have a number of advantages e They empower consumers to set the content standards depending upon their own set of values and on the age and maturity of the children involved e They provide a mechanism for consumer protection which can operate in parallel with legislative approaches or in
13. will also be sufficiently comprehensive to provide a detailed assessment of the relative quality of the tools This assessment as far as is practical and achievable will be based on quantitative measures rather than subjective qualitative measures e Standardisation In order to better inform the end users of the performance and functional characteristics of filtering software the benchmark study will provide a test 3 IDATE Prepact Review of EU Third party filtering and rating software and services Lot 3 Final Report Vol 1 Dec 1999 www idate org Benchmarking ot tiltering sottware and services Evaluation Criteria process that will become standardised across Europe This will ensure that filtering tool evaluations can be carried out in a systematic reliable repeatable and comprehensive manner Ensuring that the benchmark process is adopted as a standard approach will require a number of key activities including developing the test process within the context of known and established software evaluation standards for example ISO ensuring the acceptance and input of the producers of filtering software tools soliciting the acceptance and input of third party stakeholders such as organisations involved in consumer protection or software evaluation The benchmarking study will also investigate the development of mechanisms for the partial automation of the testing process most notably with regard to the blocking effectiveness
14. 2 7 2 1 7 2 2 Interoperability Metrics Negative Impact on the operating system System interactions co operation with other filtering tools Usability Understandability Is there any notification about a blocking action If Yes for each channel Is there an explanation of the reason for a blocking Transparency User Access to bad URL list User access to good URL list User access to KEYWORDS list User access to PICS parameters Is there a trace log file kept of the filter activity Is there a trace log file kept of the browser session activity Are log files analysable Are log files printable Friendliness Is there a default installation configuration If Yes For each channel Is there a configuration wizard Is there a User manual paper online Completeness of user Poor documentation Good Excellent Understandability of the user Poor documentation Good Excellent Is there a multimedia tutorial Is there a contextual help Is there an online F A Q list Is there customer support Mail Telephone freephone etc Is there local technical support US European single country Number of click needed to access to the functions add new URL to list download new URL list etc Understandability of the user Poor interface Good Excellent Is there a multimedia tutorial Accessibility to functions through menu icons keyboard shortcuts etc Effort required to learn how to Easy op
15. EXT a 4 PREVIOUS WORK Givi ccccscccctacssescsnsssesccscccsssestesthecesesbocsanseetecsscssscssesettesasesssccesssesessssesesosecedesseeseeesesses 4 THE BENCHMARK PROPOSAL eisissceessssccssnassccsscdesccdssnsenccnacesecasnasaccseccesnddenscoacesscdescocessssdeseenancs ences 5 THE BENCHMARK PROCESS sccccsssssssssscssscssssssssscccsssssssscccsssssssssssssssccsssssesssssesssssssssssssnses 6 4 1 EVALUATION CRITERIA 6 4 1 1 Blocking Effectiveness ie ceticeddetnsol ccudevgacodeuedivesdises ieni Ni nii Viinien rin 6 4 1 2 Over blocking Sensitivity ccccccccsscccsceeseeesssceceseceseecseecssecseecsseaecesseeesaeecseessseecssaseseeeeesaes 7 4 1 3 Security Integrity 7 4 1 4 Operational Ine Brit y ELON 7 4 1 5 COMPU SUN ADIL Ss igs ashe Sons ica Na ia Se hea oth a ote Mea 7 4 1 6 CUSTOMISED ILIDY os cess cob cs cet Shes teehee ie Hie os Mes Ni eave ens eats Wis ee ea eh one Mate ets eae 7 4 1 7 Usability a hac CEs OB ahaa Ca ha au oes 7 4 2 PEST IMETHOD OSERE 7 4 3 DATA PROGESSING METHOD EAE ONEEN 7 4 3 1 The FEM Modeli RLS CI as O66 AN AKE EE EAE ER 7 CONSTRAINTS AND REQUIREMENTS FO
16. Esprit funded project It is based on three main principles Benchmarking ot tiltering sottware and services Evaluation Criteria e Hierarchy each top level attribute for example reusability portability etc represents a top level node in a quality tree that facilitates the management of metrics complexity e Normalisation Each intermediate node has a representative value that is normalised in the range 0 1 with zero meaning worst case one meaning optimal value and 0 5 representing the acceptance threshold This normalisation allows a quick evaluation of a composite metric indicator an FCM node e Synthesis Each FCM node is represented graphically with a kiviat diagram radar diagram permitting a quick overview of the value of the node the bigger the area the better the value and of its components An example of a kiviat diagram is shown below figure1 Figure 1 Kiviat radar diagram showing presentation of three measurement metrics 5 Constraints and requirements for the benchmark process In the following section the requirements and operating constraints for the benchmark process are discussed 5 1 Defining a measure of Blocking Effectiveness The most critical performance factor for filtering tools is the effectiveness of blocking harmful content Consequently the benchmark process will define a stringent test method for evaluating filtering performance There are many factors that contribute to
17. R THE BENCHMARK PROCESS 8 5 1 DEFINING A MEASURE OF BLOCKING EFFECTIVENESS ccccceseseeeeeeeeeeeeeeceeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeees 8 5 1 1 Filtering Tool 8 5 1 2 Data Test Set Neutrality iirinn i i E EE NEE EN AN 10 5 1 3 Test Set S crecy i Ke EE EEEE NOS AE ESTEE E E 10 5 1 4 Data Test S t Stability aa Ne e A ER a EE EN Ea 10 5 1 5 Cultural and linguistic diversity cccccscceescccescecessecesseeesseecsceceseecesseceseecseessseecesaeeeseessaeees 11 5 2 EVALUATION OF USABILITY RELATED PROPERTIES 11 FILTERING SOFTWARE QUALITY ASSESSMENT 12 6 1 1 ISO 9126 Software Product 12 6 2 THE FCM QUALITY MODEL ISO 9216 DERIVED cccccccccssesssseeeeeeeeeesssneeeeceesesenssaeeeeeeeeeeeeenaea 13 6 3 SOFTWARE QUALITY FACTORS AND CRITERIA sccccccecessssssecccecceesesssnneeeececeessssaeeceeceeeesssnaeeeeees 14 6 3 1 T 15 6 3 2 Sonatas 15 6 3 3 15 6 3 4 35 T TT 16 DEFINITION OF THE MEASUREMENT METRIC ccccsssssssssssscsssssssssscccsssssssssssscessssssees 16 7 1 BUNGTIONALIT YS
18. aspects including blocking effectiveness functionality and usability the relative strengths and weaknesses of filtering tools can be identified As the demand for these tools increases driven by increasing Internet access by minors and as a result of imposed local and national legislation it is vital that the weaknesses in the current product base are identified and that both the consumer market and the producers are made aware of the limitations It is anticipated that the identification of the weaknesses will spur technical development refinement of the tools The benchmark process will also provide a means of testing that a product meets a certain minimum required level of performance especially in regard to blocking effectiveness Quality Assessment Published tests and feedback from user surveys again reference the IDATE report has firmly established that quality issues are a major concern for users of filtering software Consequently a major aspect of the benchmark process will be quality assessment that is determining the degree to which the tools fit their intended purpose Quality assessment will address many different functional and non functional properties of the filtering tools including user interface issues installation and maintainability issues and reliability and security issues A range of measurement metrics will be defined that will be sufficiently generic to cover a broad class of filtering tools but which
19. ated for the trials An alternative approach is to test each of the techniques within a tool separately and in different sets of combinations However due to the complexity of this approach it will not be adopted for the initial development trials of the benchmark test CyberPatrol CYBERsitter Safexplorer NetNanny WE CYBER 4 4 0 BLOCKER SENTINEL network Stand alone system Protocol system Filtering Technology Rating System Keyword blocking xX Protocol blocking Host blocking X X X X Characterised CR Time Restriction xX Host Client side Server side ISP s Table 1 Comparison of filtering techniques offered by six common software filtering tools Having defined the default configuration it is necessary to assess the other configuration parameters that influence the filtering performance of the tools In this respect there are a number of issues which have to be clarified Which Lists The filtering performance of the tools is largely defined by the contents of the URL lists contained within the tool Most tools contain two different types of lists which are typically described as blocked URL lists the database of prohibited URL s stored within the tool also known as black lists No lists Bad lists etc and approved URL lists the database of acceptable sites also known as white lists Yes lists
20. been typically carried out through hands on evaluation of the performance in accessing or blocking a random selection of unacceptable websites For example the free speech organisation which carried out a number of trials on filtering software created a test composed from the first 1000 alphabetically listed com URLs although an arbitrary choice this list in fact contained URL s of over 300 web sites containing harmful content The Which Organisation in their trials of May 2000 tested the software on a Set of 23 sites considered to contain harmful content Testing of the software on an arbitrary list of URL s can provide some qualitative feel for the relative performance of the tools but it is not a specific measure that can be used in a standardised benchmarking process Also the use of a small test set relative to the number of available web sites does not preclude the possibility of the test set being inadvertently biased in favour of the URL blocking lists of one or more of the tools This could possibly lead to one or more tools having an apparent performance advantage over the others Consequently it will be necessary to define a data test set that has broad URL coverage to decrease the statistical possibility of inherent bias towards an individual tool 5 1 3 Data Test Set Secrecy If a database of URL s is used to test the filtering performance of the tools it is essential that the contents of the te
21. d Netscape MS Internet Explorer Opera Others Length of time tool has been commercially available Yes No Yes No Yes No Yes No n Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No time Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No time ann Benchmarking ot tiltering sottware and services Evaluation Criteria Filtering Algorithms 7 3 3 Security Standard Advanced State of the Art e g image analysis Administrator s access password protected Support for hardware security devices e g smartcards How easily can the password be disabled How easily can the filtering software can be made disabled How easily can the filtering software be removed 7 4 Effectiveness The effectiveness metrics are performance measures Consequently they do not have simple measurement metrics as defined for the above quality criteria The metrics will be derived from Basic PC knowledge easy Normal PC knowledge moderate Expert PC knowledge hard Basic PC knowledge easy Normal PC knowledge moderate Expert PC knowledge hard Basic PC knowledge easy Normal PC knowledge moderate Expert PC knowledge hard Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No active performance evaluations as discussed within section 5 of this report D
22. erate the application Moderate Hard Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Benchmarking ot tiltering sottware and services Evaluation Criteria 7 2 3 Resource Requirements 7 2 4 7 2 4 1 Static disk usage installation Dynamic disk usage runtime Memory usage runtime Processor Requirement Observable difference in browser response time with filtering active If yes lt 1sec lt 10 sec gt 10 sec Cost Trial version available Low 0 20 Medium 20 100 High gt 100 Cost of List updates monthly Low 0 5 Medium 0 20 High gt 20 Operability Configuration Blocking list updates Automatic Manual User defined filtering configuration ie which combination of techniques User definable URL blocking lists Remove Add Import Export User definable keyword lists Remove Add Import Export Define access privileges for each URL list Predefined URL Grouping categorisation Personalised lists definition User defined URL Grouping categorisation Create User profiles Define User privileges Define User access time schedule User definable interaction with other applications Blocking of Port Numbers n n n type Yes No Yes No Yes No Y
23. es No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Yes No Benchmarking ot tiltering sottware and services Evaluation Criteria 7 2 4 2 7 3 7 3 1 7 3 2 Installability Installation CD ROM Internet download Auto installation available Installation Wizard for User System parameters Number of installation steps On line help for installation Uninstall option Can user easily re try setup installation of software Can user or maintainer easily install software to operation environment Easiness of manual install operation very easy only user s watching except just start install or setup functions easy only user s answering to question from install or setup functions not so easy user s looking up parameters from tables or fill in boxes to be changed and setting them complicated user s seeking parameter files looking up parameters from files to be changed and writing them Reliability Stability Possible to install tool with other filtering applications Interoperability and compliance with other applications non browser Meantime between errors Browsers supported Netscape MS Internet Explorer Opera Others Uninstall facility Maturity OS supported Win 95 Win 98 Win 2000 me Win NT Unix Linux Macintosh OS Solaris Browsers supporte
24. etailed discussions of the effectiveness metrics will be issued in a subsequent report that defines the test method 8 Conclusions This report has presented a first draft of a set of evaluation criteria and quality factors for the analysis and benchmarking of filtering tools and services The document has also discussed requirements relating to the design of the analysis framework and performance metrics Additionally it has described the context of the evaluation process within an internationally recognised ISO standard for software product evaluation Comments and feedback in regard to the proposed evaluation framework are welcomed This can be achieved by direct contact with the author or via the web site established to support the study http efilter jrc it This site hosts an interactive discussion forum as well as acting as an on line repository for documents relating to the study na
25. gh the Net Which Associaiton Filtering tool Evaluation May 2000 http www iwt org uk safe which total htm Benchmarking ot tiltering sottware and services Evaluation Criteria 5 1 5 Cultural and linguistic diversity The IDATE study reported that the broad range of commercially available software tools do not support the cultural and linguistic diversity of Europe The tools are largely pre configured to meet the requirements of English speaking users Clearly this poses problems for a benchmark process that is intended to be pan European However this study will proceed on the assumption that the implementation of the benchmark process will be independent of language issues and cultural issues The initial implementation will deploy testing only with respect to the English language The method will be open to be adaptated for implementation to suit any cultural or language base 5 2 Evaluation of usability related properties A benchmark is typically used to measure the physical performance attributes of an item against a predetermined reference point or mark for example the established Business Winstone metric to assess the computational performance of a CPU However when dealing with software tools and services it is typically the functional properties that are under assessment rather than physical properties The evaluation criteria that have been defined cover a range of and functional and non functional properties
26. ity of a tool to monitor log or trace the browsing activity of a user during an on line session 4 2 Test Method The test method is in the process of development It is highly dependent upon the definition of the measurement metrics As such the details of the test method will be elaborated in a subsequent report after an exercise to validate the measurement framework and after initial trials have been carried out on a sample of filtering tools 43 Data Processing Method In section 4 1 the set of evaluation factors were outlined These measurement criteria will be subsequently broken down into subcategories of measurement metrics Ref section 7 of this document An important feature of the benchmark process is the means by which the measurement metrics can be rationalised and interpreted because the test will involve an extensive range of measurements It has been decided that the benchmark process will use a technique called the FCM model Factor Criteria Metrics This method of data representation and visualisation provides a mechanism to structure the measurement data to normalise and weight the measurement data and to synthesise the display of measurement data across the metric set using kiviat diagrams radar diagrams 4 31 FCM Model The FCM Factor Metrics is a data representation model defined for the organisation and management of software metrics This model was proposed and developed under REBOOT an
27. ncy of failures due to software errors 6 3 2 2 F2C2 Stability Software attributes allowing the application to keep a specified performance level in case of software error 6 3 2 3 F2C3 Security Software attributes that bear upon the ability of the application to prevent unauthorised access to data or programs either accidental or deliberate Note that these criteria could equally well be placed under the functionality heading 6 3 3 Usability The term usability refers to the attribute of user friendliness of a piece of software User friendliness is something of a subjective concept but in general terms ti refers to the ease with which a user can interact with a tool and gain operational access to totally exploit its functionality The term users typically refers to the direct operating agent of an interactive software application Therefore users can be terminal operators final users or in direct users The term can also refer to all those who are under the influence of or depend on the use of the application 6 3 3 1 F3C1 Understandability Software attributes affecting the effort made by the user in understanding the logical concepts of the software and its functionality ar Benchmarking ot tiltering sottware and services Evaluation Criteria 6 3 3 2 F3C2 Resource Requirements Software attributes relating to the resources required to operate the software within the host computer environment
28. nformance Replaceability Efficiency Maintainability Portability Table 2 The Quality subcharacteristics defined by ISO IEC 9126 6 2 The FCM Quality Model ISO 9216 Derived The quality framework proposed under ISO 9126 is a generic model for the analysis of diverse types of software product However for the purpose of the benchmarking study it is has been decided to further simply the model and to introduce a new category that is specific to the benchmark requirements The new framework which will also be set within the FCM model discussed in section 4 3 1 is outlined in table 3 There are four principal quality factors Functionality Reliability Usability and Effectiveness Effectiveness is the category that has been introduced into the 1509126 framework for the purpose of the study ISO 9126 also provides annexes that recommend measurement metrics that are associated with each subcharacteristic However metrics have been uniquely defined for the benchmark process and are defined in section 7 Benchmarking ot tiltering sottware and services Evaluation Criteria Attribute Factor Criteria Usefulness Functionality Flexibility Interoperability Maturity Reliability Stability Security Quality Understandability Resource Requirements Usabilit Friendliness Operability Blocking Performance
29. rity Security Integrity is the term that will be used to describe the capability of the tool to prevent the filtering services being by passed by for example mechanisms to defeat the password protection or techniques to alter or remove the filtering parameters 4 1 4 Operational Integrity Operational integrity we define as an assessment of a tool s stability in use that is both its reliability in use and its effect on the reliability of other browser tools 4 1 5 Configurability Configurability is an assessment of tool s flexibility in combining diverse filtering methods for example URL blocking and or text matching according to user requirements 4 1 6 Customisability Customisability we define as the degree to which the software filters can be customised or modified according to user preferences 4 1 7 Usability Usability we define as a measure of the ease of use of the tool both from the perspective of installing and maintaining the tool and in regard to the ease with which it can be deployed during an on line browsing session Usability will also address issues such as the transparency of the tool cost effectiveness and traceability Transparency we define as the accessibility of the filtering rules or lists that are used by the service or tool to block access to web sites Cost effectiveness is a measure of the price to performance ratio of the filtering service or tool We define traceability as the capabil
30. st remain undisclosed for the period that the benchmark is in force If the contents of the test set are known then it creates the possibility for software suppliers to tailor the configuration of their software e g by adding the test data to the default blocking lists to perform well on the test set Consequently the list should only be distributed to trusted third parties or alternatively it should be securely encrypted and accessed during the benchmark trials via a deciphering script for example via a JAVA routine 5 1 4 Data Test Set Stability The data test set must remain consistent and stable for the duration that the benchmark is in force This presents a problem for test sets that are based on the URL s of active web sites The world wide web is a highly dynamic environment and web sites are liable to change close or move over short periods of time The benchmark data test set must be independent of changes in web sites There are two possible solutions to this problem Either through the use of a virtual database of URL s that remains static regardless of the actual content of the physical web site Alternatively the benchmark process may be deployed with dynamically created test sets that are valid only at the time of testing The latter option would render it more difficult to standardise the benchmark Patrol first 1000 com domains html Throu
31. the configuration of keyword lists also applies to the problem of defining domain lists Some of the filtering tools allow users to define domains e g com org that can be blocked The benchmark trials will be carried out on the assumption that domain blocking will not be activated during the evaluations Benchmarking ot tiltering sottware and services Evaluation Criteria e What Age User The typical end users of web browsers enabled with filtering software are children either at home or in public amenities such as schools or libraries However the criteria for blocking harmful context is highly age sensitive material that might be considered harmful or undesirable for young children may be perfectly acceptable for children of teenage years The benchmark process must be independent of age factors This may require different test sets configured to model the diverse age related requirements of the users Alternatively the test process may operate on the assumption of the highest required filtering scenario and test for blocking of content that would be considered harmful to the youngest age range This latter approach will be adopted for the initial benchmark trials due to time limitations on the study However it is acknowledged that end users may find a series of trials organised into an arbitrary age hierarchy more satisfactory 5 1 2 Data Test Set Neutrality Testing of blocking performance of filtering tools and services has
32. the overall filtering performance of a tool and which must be considered during the definition of the test method There are also benchmark process constraints that will define the structure and content of the data test set The issues that will play a significant factor in the definition of the test method include the following e Filtering Tool configuration e Data Test Set secrecy e Data Test Set stability e Data Test Set neutrality e Cultural and linguistic diversity 5 1 1 Filtering Tool Configuration Commercially available filtering tools deploy a wide range of techniques for blocking internet content The most common of these are URL blocking lists or approved lists text or key word matching and PICS label scheme In order to ensure consistency of the benchmarking trials it is necessary to define a common basepoint or default configuration for the tools This is complicated by the fact that the tools do not contain the same combinations of available techniques and may combine URL filtering lists with one or more of the other methods Table 1 Benchmarking ot tiltering sottware and services Evaluation Criteria lists the comparative techniques and methods for cross section of common filtering tools For the initial trials of the test process we have adopted the convention that the tools should be tested and configured at their maximum capability that is with all of the available techniques within a tool a specific tool activ
33. the situations where enforcement of law fails can operate independently as a self imposed monitoring scheme However the performance of these tools both in terms of usability and effectiveness is variable and often does not meet the requirements of the individual users 2 Previous Work Evaluations of filtering software have been carried by many diverse organisations including government academia consumer organisations freedom of speech groups and in the traditional computer 2 Although these studies have been conducted for different social or scientific reason and have taken differing technical approaches they have consistently highlighted the 1 Concepts Document Benchmarking of Filtering Software and Services Jackson Oct 2000 http e filter jrc it For example Access Prevention Techniques for Internet Content Filtering P Greenfield P McCrea S Ra National Australian Office for the Information Economy http www noie gov au publications NOIE CSIROfinalreport htm Through the Net Which Association Filtering Test May 2000 http www iwf org uk safe which total htm lt x html gt Why Internet Content Rating and Selection does not work Kristian and Marit K6hntopp http Awww koehntopp de kris artikel rating does not_work Canada Study on rating and filtering Industry Canada http strategis ic gc ca SSG it05082e html Reviews of Internet Access Filtering Software SuperKids http Awww
34. tionality A Software application is held to be functional to the extent that the procedures it contains coincide with the functions required In other words the functionality refers to the compliance of the application with the user s expectations and requirements 6 3 1 1 F1C1 Usefulness Software attributes affecting the presence and the adequacy of all functions for a specific target 6 3 1 2 F1C2 Accuracy Software attributes concerning the generation of correct results or actions 6 3 1 3 F1C3 Interoperability Software attributes affecting the capacity to interact with specific systems 6 3 2 2 Reliability The traditional definition of software reliability refers to an application s ability to maintain its specified performance whilst operating under specific conditions and for a specific period of time For the purposes of the benchmark study we broaden the scope of this definition to include other performance variables such as security and the ability of the software to interact correctly with other unspecified software applications Reliability in the context of filtering tools refers to the users ability to operate the software within their host computer environment with a high degree of confidence that is going to fulfil all its functional requirements whilst operating with stability and a sufficient level of security protection against misuse 6 3 2 1 F2C1 Maturity Software attributes affecting the freque
35. ustomisability e Usability These measurement factors encapsulate the broad range of functional and non functional properties of the tools that are the major areas of concern for end users In the following sections each of these parameters will be defined 4141 Blocking Effectiveness The measure of blocking effectiveness we define as the relative performance of the tool in blocking harmful content That is to what degree is the tool successful in preventing harmful internet content being displayed within a browser during an on line Internet session Clearly this will be one of the most important measures applied to the filtering tools If possible the analysis should be quantitative rather than qualitative See the EU report on Internet content rating for a description of what has been defined as harmful content within the study COM 96 487 Final Illegal and Harmful content on the Internet Communication from the EC Green paper on the protection of Minors and Human Dignity in Audio visual and information services Benchmarking ot tiltering sottware and services Evaluation Criteria 41 2 Over blocking Sensitivity Over blocking we define as the measure of the tools to tendency to incorrectly block access to legitimate web sites that do not contain harmful content It should be noted that over blocking has been consistently recognised as one of the major weaknesses of filtering software tools 41 3 Security Integ

Download Pdf Manuals

image

Related Search

Related Contents

PA2015017 - Banque de France  カロラ島専用 PG-ー 。 取扱説明書  Istruzioni di montaggio PYROMAT ECO  Operator`s Manual  Samsung PL51 Užívateľská príručka  RBT-001ユーザーズガイド  MANUAL DE INSTRUÇÕES Profibus-DP Junction Box DP    

Copyright © All rights reserved.
Failed to retrieve file