SPECsfs2008 User’s Guide

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Standard Performance Evaluation Corporation (SPEC)

                                 6585 Merchant Place, Suite 100

                                 Warrenton, VA 20187, USA

Phone: 540-349-7878

                                 Fax:     540-349-5992

                                 E-Mail:    info@spec.org

www.spec.org

 

 

 

 

 

Copyright (c) 2008 by Standard Performance Evaluation Corporation (SPEC)

All rights reserved

SPEC and SFS are registered trademarks of the Standard Performance Evaluation Corporation

NFS is a registered trademark of Sun Microsystems, Inc.



 

Table of Contents

 

1     Quick Start Guide. 5

1.1        Prerequisites. 5

1.2        Installing SPECsfs2008. 6

1.3        Editing the configuration file on the Prime client 7

1.4        Configuring the server for testing. 8

1.5        Starting the benchmark. 8

1.6        Monitoring the benchmark execution. 8

1.7        Examining the results after the benchmark execution has completed. 9

2     Introduction. 10

2.1        What is new in SPECsfs2008. 10

2.2        SPECsfs2008 Benchmark Overview.. 10

2.2.1      NFS Workload. 11

2.2.2      CIFS Workload. 12

2.2.3      Comparing the NFS and CIFS Workloads. 13

3     Installing and Configuring the Benchmark Environment 14

3.1        Setting up the System Under Test (SUT) 14

3.2        Setting up the Load Generators. 15

3.2.1      Configuring SFS2008 Windows Clients for Auto-Startup. 16

3.3        Configuring the Required Benchmark Parameters. 16

3.3.1      Other Variables in the RC File. 18

4     Running the Benchmark and Interpreting Results. 20

4.1        SFS Benchmark Directory Structure. 20

4.2        Pre-Compiled SFS Benchmark Binaries. 20

4.3        Using the SFS Manager 21

4.3.1      Example of SUT Validation. 22

4.3.2      Example of a Benchmark Run. 23

4.4        Interpreting the Benchmark Results. 24

5     Submission and Review Process. 26

5.1        Creating Reports. 26

5.1.1      Creating the Submission File. 26

5.1.2      Creating the Submission Reports. 29

5.1.3      Submit Tools Example Files. 30

5.1.4      Submit Tools Example Scenarios. 31

5.1.5      Translation of Free-Form Fields. 39

5.2        Submitting Results. 40

6     FAQ.. 41

6.1        SPECsfs2008 Benchmark Press Release. 41

6.2        Running and troubleshooting the benchmark. 44

6.3        Tuning the Server 45

6.4        Submission of Results. 46

7     Bibliography. 47

8     Trademarks. 48

9     Appendix A – Building SFS Benchmark Components. 49

9.1        Building SFS for UNIX.. 49

9.2        Building SFS for Windows. 49

9.2.1      Update Visual Studio Libraries/Includes. 49

9.2.2      Open the Visual Studio workspace files. 49

9.2.3      Build the individual project files. 49

9.2.4      Debugging with Visual Studio. 50

9.3        Building the SFS Manager 50

9.4        Building the Result Submission Tools. 50

10       Appendix B – Setting up password-less ssh. 52

 


1         Quick Start Guide

 

The SPECsfs2008 benchmark is used to measure the performance of an NFS or CIFS server. It creates a mixed workload that simulates a typical server environment. The user may use the tool to measure the performance of an NFS or CIFS server and compare the results with other servers that have results published on the SPEC web site.

 

SPECsfs2008 results summarize the server’s capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. A minimal SPECsfs2008 result will contain ten load points and the results of the server at each of the requested load points. Each load point presents a requested number of operations per second.

 

The benchmark requires that the user have one or more client systems that can be used to generate load on the server. One of these clients, referred to as the ‘prime’ client, is used to coordinate all the load-generating clients (the prime client itself is eligible, but not required, to generate load).

 

Example topology

 

 

The minimal configuration consists of one load generating client, and one NFS or CIFS server. The server is often referred to as the SUT (Server Under Test) in the documentation below.

 

The steps to produce a SPECsfs2008 result are:

 

·         Install SPECsfs2008 on the load generators

·         Edit the configuration file on the prime client

·         Configure the server for testing

·         Start the benchmark

 

 

1.1        Prerequisites

 

·         The Java RunTime Environment (jre) release 1.5 or later must be installed on the load generator (client) and Prime Client systems.

·         The J2SE(TM) Development Kit release 5.0 or later must be installed on the system(s) where you want to build the java code in the benchmark (not normally required)

·         The correct java program must be in the current user’s path.

·         The test file systems (exported via NFS V3 or as CIFS shares) must have the permissions set correctly in order to allow access and read/write operations by the clients.  The test file systems do not need to be mounted or mapped prior to execution of the benchmark.

·         There must be network connectivity between the SUT (System Under Test) and clients, and between the clients and the Prime Client.  The Prime Client is simply the system on which the benchmark run is started, and could be one of the clients or the SUT.

·         The contents of the SPECsfs2008 benchmark CD must be accessible on all the systems where the benchmark will be installed.

·         Using these quick start procedures assumes that the pre-compiled C code binaries shipped with the benchmark will be used.

 

1.2        Installing SPECsfs2008

 

SPECsfs2008 can be installed on client machines running either a UNIX-based or Windows operating system. Each of these require slightly different configuration and are described separately below.

 

UNIX client installation and configuration:

·         Ensure that DNS is correctly configured.

·         Install Java 1.5 or later, and ensure that java is in the user’s search path. UNIX versions of Java may be downloaded from http://www.java.com.

·         Install SPECsfs2008 benchmark using the following steps:

o        Login to the client (as root)

o        Insert the SPEC CD-ROM and mount the CD-ROM device.

o        cd to the top level CD-ROM directory

o        Enter ‘cd spec-sfs2008/manager’

o        Enter  ‘java SpecInstall “destination_directory”’ (where “destination_directory“, enclosed by double-quotes, is where you wish to have the benchmark installed)

 

Windows client installation and configuration:

·         Ensure that Windows Internet Naming Service (WINS) is correctly configured.

·         Install Java 1.5 or later, and ensure that java is in the user’s search path. Windows versions of Java may be downloaded from http://www.java.com

·         Install SPECsfs2008 benchmark

o        Start a command prompt window (DOS type command shell). This can be done using the ‘Start’ button, choosing ‘Run…’ and entering ‘cmd’.

o        Insert the SFSsfs2008 CD-ROM and enter the drive letter associated with the CD-ROM device (eg. D:) at the command prompt.

o        Enter ‘chdir spec-sfs2008\manager’

o        Enter ‘java SpecInstall ‘“destination_directory”’ (where “destination_directory”, enclosed by double-quotes, is where you wish to have the benchmark installed)

 

Note: If one is running SPECsfs2008 on a Windows workstation client. (non-server version of Windows) then one must disable the Windows  file server component before using the client to run the SPECsfs2008 benchmark. This is accomplished by using a command window and running:

 

                                net stop server

 

This applies to all non-server versions of Windows (NT, XP, Win2k, and Vista )

 

1.3        Editing the configuration file on the Prime client

 

On the Prime client, copy one of the sfs ‘rc file’ templates (sfs_nfs_rc for NFS, sfs_cifs_rc for CIFS)  in the manager directory into a file called sfs_rc. The user must edit the sfs_rc configuration file, and only needs to edit it on one of the clients. The client that contains the edited configuration file must be the Prime client.  The user does not need to edit, or even have, a configuration file on the other load generating clients.  On the Prime client, edit the values for:

 

o        LOAD                                           
Initial value for requested operations/sec, or a complete list of the data points to be collected by the benchmark. Examples are provided in the sfs_nfs_rc and sfs_cifs_rc template files, which are located in the manager directory.

 

o        INCR_LOAD                              
Incremental increase in load for successive data points in a run.  This parameter is used only if LOAD consists of a single (initial) value.  To ensure equally spaced points, the value of LOAD and INCR_LOAD must be equal.

 

o        NUM_RUNS                
The number of load points to run and measure (minimum of 10 for a publishable result).  This parameter is used only if INCR_LOAD is specified.

 

o        PROCS                                        
Number of processes per client. Each client load generator may be able to generate more load if the client has sufficient resources to do so.  A general rule of thumb is to have the total requested load be divided across all of the clients, and to have sufficient numbers of clients and processes so as to have the operations/sec per process remain below 250 at the highest load point.  It is also recommended to have the operations/sec per process remain above 10 at the lowest load point.

 

o        CLIENTS                                     
List of clients to use in this test. The Prime client, if listed here, may also be used to generate load. If the Prime client is not listed here then it will only coordinate the testing and will not participate in generating load.  The client names in this list are hostnames or IP addresses of the clients that will be participating in generating the load.

 

o        MNT_POINTS

List of mount points, or shares, to use in the testing. Each of these mount points must be exported by the server so that they may be mounted by the load generating clients. The value MNT_POINTS can take several different forms:

§         UNIX style:  server:/exportfs1 server:/exportfs2 …

§         CIFS style: \\server\exportfs1  \\server\exportfs2

§         Use a file that contains the mount points:  filename
The use of a file, and its format, is covered later in this document.

The number of mount points in the list must be equal to number of processes specified in the PROCS parameter.  Note that a mount point may be repeated in the list.

 

o        FS_PROTOCOL

The type of server protocol (NFS or CIFS) to test.  It may be set to “nfs” or “cifs”. Either UNIX or Windows clients can be used to test either NFS or CIFS, however, all clients must be of the same type. Note:  If this value is set to “nfs” then the MNT_POINTS list must use the UNIX style syntax. If this value is set to “cifs” then the MNT_POINTS list must use the CIFS style syntax.

 

When testing a CIFS configuration:

o        USERNAME                 Specify the CIFS user name for accessing the server.

o        PASSWORD                               Specify the CIFS password for accessing the server.

o        DOMAIN                                       Specify the CIFS domain for accessing the server.

 

When testing an NFS configuration from a Windows client:

 

o        SFS_NFS_USER_ID                  Specify the NFS server user ID to be used.

o        SFS_NFS_GROUP_ID              Specify the NFS server group ID to be used.

 

1.4        Configuring the server for testing

 

·         Export or share file systems that will be used for testing (export for UNIX, or enable Shared folders for Windows). The exported name must match the values specified in the MNT_POINTS variable in the SPECsfs2008 configuration file.

·         Ensure the exported file systems have read/write permissions.

·         Ensure access is permitted for username, password, and domain. (CIFS testing only)

 

1.5        Starting the benchmark

 

Note that the SfsManager must be run under the same user id (UID) on the all of the clients, including the prime client.

 

·         Change directories to the destination_directory specified during install.

·         On the Prime client:

o        Enter ‘cd manager’

o        Enter ‘java SfsManager –r sfs_config_file –s output_files_suffix –v 2’
(This performs the server validation)

o        Enter ‘java SfsManager –r sfs_config_file –s output_files_suffix’
(This runs the actual test)

 

1.6        Monitoring the benchmark execution

 

·         On the Prime client, change directories to the destination_directory from the installation step above by entering ‘cd result’

 

The user may now examine the benchmark logs, as well as the results. As the benchmark runs, the results are stored in the files with names like:

sfssum.*                                Summary file used in the submission process described later.

sfslog.*                                  Log file of the current activity.

sfsres.*                                  Log file of the current activity.

               

During the run, each client will store its client log file ‘sfsc_log’ in its local /tmp directory. After all load points are complete, the files from each client are collected into the result directory on prime client and renamed with the client number. The client logs are files with names like: sfsc0001.*

 

sfsc*.*                                   The client log files.

 

1.7        Examining the results after the benchmark execution has completed

 

The results of the benchmark are summarized in the sfssum.* file in the result directory on the prime client.  This may be examined with any text editing software package. This file is the summary file that may be used for the submission process, described later in this document.

 


2         Introduction

 


SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation benchmark that measures file server throughput and response time. It provides a standardized method for comparing performance across different vendor platforms. 

 

This document specifies how SPECsfs2008 is to be run for measuring and publicly reporting performance results, and includes a guide to using the SFS tools.  The SPECsfs2008 Run and Reporting Rules (included in a separate companion document on the SPECsfs2008 CD) have been established by the SPEC SFS Subcommittee and approved by the SPEC Open Systems Steering Committee. They ensure that results generated with this suite are meaningful, comparable to other generated results, and are repeatable. Per the SPEC license agreement, all results publicly disclosed must adhere to these Run and Reporting Rules.
       
SPEC requires that any public use of results from this benchmark follow the
SPEC OSG Fair Use Policy.  In the case where it appears that these guidelines have not been adhered to, SPEC may investigate and request that the published material be corrected.

 

The section below describes the changes in SPECsfs2008 compared to the previous version of the benchmark, SFS 3.0. 

2.1        What is new in SPECsfs2008

 

The SPECsfs2008 release of the benchmark includes major workload and functionality changes, as well as clarification of run rules.  The code changes compared to earlier SFS versions were NOT performance neutral, therefore comparing SPECsfs2008 results with SFS 3.0 results is NOT allowed.

 

The most significant areas of change in SPECsfs2008 include:

 

  • An updated NFSv3 workload, based on recent data collected by SFS committee members from thousands of real NFS servers operating at customer sites.
  • The addition of a general-purpose CIFS fileserver workload, also developed using data collected from thousands of real CIFS servers.
  • Removal of support for the NFSv2 and support for the UDP transport.
  • Support for Windows and MacOSX clients (for generating either NFS or CIFS workloads).
  • A new java-based SFS manager test harness.
  • A new XML based reporting and submission framework.
  • Simplified and better organized documentation.
  • Significant improvements to benchmark stability and error handling.

 

2.2        SPECsfs2008 Benchmark Overview

 

The SPECsfs2008 benchmark is used to test the performance capacity of NFS and CIFS fileservers. Performance is measured in both in terms of throughput (the number of operations the server can serve) as well as the response time (time required to complete individual operations).

 

In a typical SFS test configuration, a series of load generating clients are directed through a network at file systems shared or exported from the ‘server under test’ or SUT. Clients are directed by the SFS manager program to execute a series of ten or more tests at increasing, equally spaced ‘load points’. A load point represents a throughput level (number of operations) for the workload the SFS clients will present to the server. At the end of each load point, the clients will report how much throughput was actually achieved, as well as the average response time for all requests presented to the server during the measurement interval. The benchmark reporting tools can be used after the full run is complete to present the peak achieved throughput as well as calculate an ‘overall response time’ metric, which is meant to reflect the results of average response time measurements across the series of load points (essentially it is the area under the throughput vs. average response time curve).

 

At the start of each load point each client will start a specified number of load generating processes. Each process will mount the exported or shared file systems, create a directory structure, and fill it with a series of files of various sizes. The number of files the load generating process creates is determined by and increases proportionally with the number of operations per second specified for the given load point.

 

When all load generating processes have completed the initialization of the fileset (known as the ‘INIT’ phase of the benchmark), they will begin to request from the server a series of NFS or CIFS operations based on the workload parameters established by the benchmark. The load generators will spend 300 seconds running in WARMUP mode, where no official measurements are recorded. After the WARMUP phase is completed, the benchmark will start a 300 second measurement phase, where the actual measurement of throughput and response time for the load point is collected.

 

Upon successful completion of the 300 second measurement period, statistics from all load generators on all clients are collected and centralized by the SFS manager program, which will report and record the overall results of the test. If the test completes without violating any benchmark rules, the summary report produced by the SFS manager may be used to produce an SFS submission which can be sent to SPEC for review and publication.

 

In the following subsections, some additional specifics regarding the NFS and CIFS workloads are provided.

2.2.1       NFS Workload

 

The NFS workload in SPECsfs2008 is intended for servers which can serve the NFS version 3 protocol over the TCP transport. Support for NFS version 2 and the UDP transport are no longer officially supported by the benchmark.

 

In many ways, the workload in SPECsfs2008 is similar to the workload presented by SFS 3.0. However, there are a number of changes that significantly alter the effect the workload is likely to have on the server under test. These changes were mainly driven by the need to adapt the original benchmark to changes that have occurred in real NFS server configurations, file sets, and workload patterns in the years since the last benchmark was released. Here is a list of the most significant changes made to the NFS workload for SPECsfs2008.

 

ü       The percentage of each NFS operation has been adjusted to better reflect more recent data collected on real NFS servers. A table later in this section summarizes the adjustments.

ü       The maximum file size was increased from 1 Megabyte to 32 Megabytes.

ü       The total file set size created for a given load point has been increased from 10 Megabytes per SFS op/s to 120 Megabytes per SFS op/s

ü       The percentage of files in the file set accessed during the test was increased from 10 percent to 30 percent.

ü       The maximum logical transfer size was increased from 135 Kilobytes, to 256 Kilobytes.

ü       The NFS block (physical transfer) size was previously fixed at 8 Kilobytes. It is now automatically negotiated with the server.

ü       COMMIT operations are no longer issued if the server returns the STABLE bit on the write operations. When a server returns an UNSTABLE response to a write operation, a COMMIT may be issued after a number of WRITE operations are issued. Previously, the benchmark would issue a COMMIT after every logical write operation was completed, regardless of whether or not the operation was STABLE. This was determined to not reflect the behavior observed by real client/server interactions.

 

The shift in NFS operations is outlined in the table below comparing the NFS operation mix in SFS 3.0 and SPECsfs2008. Note that the percentage of READ and WRITE operations in SPECsfs2008 are now expressed as ‘logical’ operations. This is because the number of physical IOs required to complete the logical read or write may vary depending on the transfer size negotiated between the SFS client and the server under test. Also note that COMMITs are no longer included in the op mix and are not counted as completed operations in the benchmark result. When required due to UNSTABLE write responses from the server, COMMITs will be issued and the time required to complete the COMMIT will be included in the response time measurement for the logical write which required it. In effect, COMMIT operations are ‘overhead’ for which no credit is given in situations when they are required by the nature of the server response.

 

NFS Version 3 Operation

SPECsfs2008

SPEC SFS 3.0

LOOKUP

24%

27%

READ

18%

18%

WRITE

10%

9%

GETATTR

26%

11%

READLINK

1%

7%

READDIR

1%

2%

CREATE

1%

1%

REMOVE

1%

1%

FSSTAT

1%

1%

SETATTR

4%

1%

READDIRPLUS

2%

9%

ACCESS

11%

7%

COMMIT

NA

5%

 

2.2.2       CIFS Workload

 

The CIFS workload in SPECsfs2008 shares the file set attributes with the NFS workload. However, the operations applied to the server under test are, of course, quite different. As with NFS, a study of thousands of CIFS servers deployed at customer sites was used to determine a target operation mix. CIFS, however, is a ‘stateful’ protocol, which implies that operations must be delivered in valid sequences. For example, a file must be opened before certain operations can be issued to the server, and closed after those operations are completed. Clearly, delivering operations in a random succession, as is done with the ‘stateless’ NFS workload, would not work.

 

To overcome the more restrictive requirements of CIFS, as well as to improve the realism of the CIFS workload, an operation generation technique utilizing a Hidden Markov Model (HMM) was developed for SPECsfs2008. Real CIFS traces collected from active CIFS servers were used to train the model, and the resulting CIFS workload reflects the valid operation sequences exhibited in the original traces.

The operation mix produced by the SPECsfs2008 CIFS workload is captured in the table below.

 

CIFS Operations

SPECsfs2008

ECHO

0.5%

LOCKING_ANDX

1.1%

NT_TRANSACT_QUERY_SECURITY_DESC

1.6%

READ_ANDX

20.5%

TRANS2_QUERY_FS_INFORMATION

1.6%

TRANS2_QUERY_FILE_INFORMATION

12.9%

TRANS2_QUERY_PATH_INFORMATION

21.5%

TRANS2_SET_FILE_INFORMATION

3.2%

TRANS2_FIND_FIRST2

8.6%

WRITE_ANDX

8.6%

NT_CREATE_ANDX

9.7%

CLOSE

9.7%

FLUSH

0.5%

 

2.2.3       Comparing the NFS and CIFS Workloads

 

While there are some similarities, especially with respect to the file sets each workload operations on, the NFS and CIFS workloads are not comparable and no conclusions about the ability of a given SUT to perform NFS versus CIFS operations should be made by comparing the NFS and CIFS results for that SUT. For example, if the CIFS results for an SUT are 20% higher than the NFS results for the same SUT, it should not be inferred that the SUT is ‘better’ at delivering CIFS operations than NFS operations. The workloads are very different and no attempt was made to normalize the NFS and CIFS workloads. The only valid comparisons that can be made are between published results for different SUTs operating against the same SPECsfs2008 workload, either NFS or CIFS.


3          Installing and Configuring the Benchmark Environment

 

This section provides information on hardware/software configuration requirements for the load generators and the file servers (CIFS and NFS).  It also includes installation instructions for the benchmark on the load generators for each of the supported operating systems.

 

3.1        Setting up the System Under Test (SUT)

 

There are several things you must set up on your server before you can successfully execute a benchmark run.

1.        Configure enough disk space. SPECsfs2008 needs 120 MB of disk space for each NFS or CIFS ops/sec you will be generating, with space for 10% growth during a typical benchmark run (10 measured load levels, 5 minutes per measured load). You may mount your test disks anywhere in your server's file space that is convenient for you. The maximum NFS or CIFS ops/sec a server can process is often limited by the number of independent disk drives configured on the server. In the past, a disk drive could generally sustain on the order of 100-200 NFS or CIFS ops/sec. This was only a rule of thumb, and this value will change as new technologies become available. However, you will need to ensure you have sufficient disks configured to sustain the load you intend to measure. 

2.        Initialize and mount all file systems. According to the Run and Reporting Rules, you must completely initialize all file systems you will be measuring before every benchmark run. On UNIX systems, this is accomplished with the “newfs” command. On a Windows system the “FORMAT” utility may be used. Just deleting all files on the test disks is not sufficient because there can be lingering effects of the old files (e.g. the size of directory files, location of inodes on the disk) which affect the performance of the server. The only way to ensure a repeatable measurement is to re-initialize all data structures on the disks between benchmark runs. However, if you are not planning on disclosing the result, you do not need to perform this step.

3.        Export or share all file systems to all clients. This gives the clients permission to mount, read, and write to your test disks. The benchmark program will fail without this permission.

4.        Verify that all RPC services work. The benchmark programs use port mapping, mount, and NFS services, or Microsoft name services, and file sharing, provided by the server. The benchmark will fail if these services do not work for all clients on all networks. If your client systems have NFS client software installed, one easy way to do this is to attempt mounting one or more of the server's exported file systems on the client.  On a Windows client one may try mapping the shares to ensure that the services are correctly configured on the CIFS server.

5.        NFS servers generally allow you to tune the number of resources to handle TCP requests.  When benchmarking using the TCP protocol, TCP support is of course required, and you must also make sure that UDP support is at least minimally configured or the benchmark will fail to initialize.

6.        Ensure your server is idle. Any other work being performed by your server is likely to perturb the measured throughput and response time. The only safe way to make a repeatable measurement is to stop all non-benchmark related processing on your server during the benchmark run.

7.        Ensure that your test network is idle. Any extra traffic on your network will make it difficult to reproduce your results, and will probably make your server look slower. The easiest thing to do is to have a separate, isolated net­work between the clients and the server during the test.  Results obtained on production net­works may not be reproducible. Furthermore, the benchmark may fail to correctly con­verge to the requested load rate and behave erratically due to varying ambient load on the network.

 

At this point, your server should be ready for a benchmark measurement. You must now set up a few things on your client systems so they can run the benchmark programs.

 

3.2        Setting up the Load Generators

 

Running the SfsManager requires that the Java Runtime Environment (JRE) release 5.0 or later be loaded on the system. You can get the JRE package from www.java.com.

 

  1. On UNIX systems, create “spec” user. SPECsfs2008 benchmark runs should be done as a non-root user.
  2. The SPECsfs2008 programs must be installed on clients.

 

To install the SPECsfs2008 programs:

 

                                On all the clients:

                                                Login as “root”

Enter ‘cd cdrom_directory/manager’

                                                Enter ‘java SpecInstall “destination_directory”’

 

  1. On UNIX systems, ensure sfsnfs3 is setUID root, if necessary. Some NFS servers only accept mount requests if sent from a reserved UDP or TCP port, and only the root user can send packets from reserved ports. Since SPECsfs2008 generally is run as a non-root user, the sfsnfs3 program must be set to execute with an effective UID of root.
  2. To get the benchmark to use a reserved port, you must include a -DRESVPORT option in your compile com­mand. This is easiest to accomplish by editing the Makefile section in which compiler options for the specific operating system used on your client systems.  The build process will then make the client use a reserved port and will arrange to run the benchmark programs as root. However, you may want to verify this works the first time you try it.
  3. Configure and verify network connectivity between all clients and server. Clients must be able to send IP packets to each other and to the server. How you configure this is system-specific and is not described in this document. Two easy ways to verify network connectivity are to use a “ping” program or the netperf benchmark (http://www.netperf.org).
  4. Before starting the benchmark, ensure that the prime client can execute commands on the remote clients using ssh with no password challenges.  Refer to Appendix B for an example of how to do this.
  5. If clients have NFS client code, verify they can mount and access server file systems. This is another good way to verify your network is properly configured. You should unmount the server's test file systems before running the benchmark. If configuring SPECsfs2008 to test CIFS, one can test that the clients can map the shares. You should un-map the shares before running the benchmark.
  6. The Prime Client must have sufficient file space in the SPECsfs2008 file tree to hold the result and log files for a run. Each run generates a log file of 10 to 100 kilobytes, plus a result file of 10 to 100 kilobytes. Each client also generates a log file of one to 10 kilobytes.

 

* IMPORTANT * -- If Windows Firewall is turned on; each program will need to be added to the exceptions list. Either open the Windows Firewall control panel and add the applications manually, or wait for the pop-up to appear after the first execution of each application. Other locally-based firewall applications may require a similar allowance.

 

3.2.1       Configuring SFS2008 Windows Clients for Auto-Startup

 

The following are the steps to follow to configure Windows clients in order to allow the Prime Client to communicate with them directly and remotely start the SfsManager process when a benchmark run is started.

 

Granting DCOM Remote Launch permissions:

 

1. Click Start, click Run, type DCOMCNFG, and then click OK.

 

2. In the Component Services dialog box, expand Component Services,

    expand Computers. 

 

3. Right mouse click on My Computer and select properties.

 

The My Computer dialog box appears. 

 

4. In the My Computer dialog box, click the COM Security tab.

 

5. Under Launch and Activate Permissions, click Edit Limits. 

 

6. In the Launch Permission dialog box, follow these steps if your

   name or your group does not appear in the Groups or user names list:

     a.  In the Launch Permission dialog box, click Add. 

     b.  In the Select Users, Computers, or Groups dialog box,

         add your name and the group in the Enter the object names to

         select box, and then click OK.

 

7. In the Launch Permission dialog box, select your user and group in

   the Group or user names box. In the Allow column under Permissions

   for User, select Remote Launch, and then click OK. 

 

3.3        Configuring the Required Benchmark Parameters

 

Once you have the clients and server configured, you must set some parameters for the benchmark itself, which you do in a file called the “rc file”. The actual name of the file is a prefix picked by you, and the suffix “_rc”. The default versions shipped with the benchmark are delivered as “sfs_nfs_rc” and “sfs_cifs_rc” in the benchmark source directory.  One may use any text editor to modify parameters in the rc files. Please note that any variable which is a list of values must have its value enclosed in double quotes.

 

There are several parameters you must set, and several others you may change to suit your needs while performing a disclosable run. There are also many other parameters you may change which change the benchmark behavior, but lead to an undisclosable run (for example, turning on debug logging).  See the SPECsfs2008 Run Rules for the classification of all the parameters.

 

The parameters you must set are:

  1. MNT_POINTS: This parameter specifies the names of the file systems the clients will use when testing the server. It can take two forms.

The first form is a list of host:path pairs specifying the file systems this particular client will be using. For example, if the server is named “testsys” and has three test mount points named “/test1”, “/test2”, and “/test3”, the list would be “testsys:/test1 testsys:/test2 testsys:/test3”, for an NFS configuration, and \\testsys\test1 \\testsys\test2 \\testsys\test3 for a CIFS configuration. You must be very careful when specifying the mount point to comply with the uniform access rule (see below). The second form is simply the name of a file containing a list of mount points for each client. The format of the file is:

 

                 For an NFS configuration:

   client_name server:path server:path...

                   client_name server:path server:path...

 

                For a CIFS configuration:

                   client_name \\server\path \\server\path…

                   client_name \\server\path \\server\path…

 

And so on, one line for each client system. This file gets stored in the “manager” directory, the same place as the rc file.

 

NOTE:  The mount points file must be stripped of carriage-return line-feeds.  For example, on Linux vi, use the “:set filemode=unix” option when editing this file.

 

  1. LOAD, INCR_LOAD, and NUM_RUNS: These parameters specify the aggregate load the clients will generate. You can specify the load points two ways:

 

• You can explicitly specify a list of load levels in LOAD. So, if you want to test a load of 100, 300, and 700 NFSops, you would set LOAD to “100 300 700".

   If you want to test a set of evenly spaced load points, set all three parameters. Set LOAD to the lowest load level, set INCR_LOAD the amount you would like to increase the load for each measured run, and set NUM_RUNS to the number of times you would like to increment the load. This is the easiest way to configure a disclosable run.

 

For example, if you would like to measure 10 evenly spaced points ending at 2000 NFSops, you would set LOAD to 200, INCR_LOAD to 200, and NUM_RUNS to 10.    If CLIENTS is 4 and NUM_PROCS is 8, note that due to rounding that occurs within the benchmark code, this will result in load points of 192 384 576 800 992 1184 1376 1600 1792 1984.

  1. CLIENTS: This is the names of all the client systems you will use to load your server. If you will be generating load with the prime client, include it on this list.
  2. NUM_PROCS: This is the number of load generating processes (“procs”) you want to run on each client system. As you add procs, you can have more operation requests outstanding at any given time, and you can use more file systems on the server, all of which tends to increase the load your server can process (until either the disks or the processors run out of capacity).

There is a relationship between the value of PROCS, CLIENTS and MNT_POINTS. The number of mount points specified in MNT_POINTS must equal the value of PROCS, or equal the value of PROCS times the number of clients in CLIENTS. In the first case, each mount point will be accessed by one proc on each client. In the second case, each listed mount point will be accessed by exactly one proc on one client. The first PROC mount points will be used by the first client, the second PROC mount points by the second client, and so forth.

You may specify the same mount point multiple times in MNT_POINTS. This allows you to have more than one process accessing a given file system on the server, without having all clients loading that file system. If a file system traverses multiple disks, (Example RAID Level  0, 1 …), then care must be taken to conform to the uniform access rule.

  1. FS_PROTOCOL:  Set this to NFS or CIFS, which are the only supported protocols.
  2. USERNAME: Set this to the user account name for the CIFS share. (CIFS testing only)
  3. PASSWORD: Set this to the user’s password for the CIFS share. (CIFS testing only)
  4. DOMAIN: Set this to the correct CIFS domain. (CIFS testing only)

 

There are many other parameters you can modify in the rc file, but generally none are necessary. They allow you to change the NFS operation mix, change run duration parameters, or turn on debugging information.

 

3.3.1       Other Variables in the RC File

 

As mentioned above, there are many more parameters you can set in the RC file. Here is the list and what they do:

 

  1. TCP: Leave it unset, or set it to 1 or “on” to use TCP to communicate between the clients and the server.  UDP is not supported.
  2. BIOD_MAX_READS and BIOD_MAX_WRITES: SPECsfs2008 emulates the read-ahead and write-behind behavior of NFS block I/O daemons. These allow a client to have multiple read and write requests outstanding at a given time. BIOD_MAX_READS and BIOD_MAX_WRITES configure how many read or write operations SPECsfs will transmit before stopping and waiting for replies. You can set these to any value from 0 to 32, inclusive. (NFS testing only)
  3. SFS_DIR and WORK_DIR: These are the directory names containing the SPECsfs2008 programs (SFS_DIR), the RC file, and logging and output files (WORK_DIR). If you configure your clients with the same path for these directories on all clients, you should not need to fool with this. One easy way to accomplish this is to export the SFS directory tree from the prime client and NFS mount it, or for CIFS map the share, at the same place on all clients.
  4. PRIME_MON_SCRIPT and PRIME_MON_ARGS: This is the name (and argument list) of a program which SPECsfs will start running during the measurement phase of the benchmark. This is often used to start some performance measurement program while the benchmark is running so you can figure out what is going on and tune your system.

Look at the script “sfs_ext_mon” in the SPECsfs2008 source directory for an example of a monitor script.

  1. WARMUP_TIME and RUNTIME: These set the duration of the warmup period and the actual measurement period of the benchmark.
  2. INIT_TIMEOUT:  The maximum time (in seconds) that the benchmark will run during the working set initialization phase for a single data point before timing out.  This value may be increased as needed, e.g., when using a slow I/O subsystem, in order to keep the benchmark from timing out during initialization.
  3. BLOCK_SIZE:  The maximum block (RPC) size which the load generators will use for network communication with the NFS server.  If this value is not set, the load generators will auto-negotiate the block size with the server based on the server’s advertised preferred size.  (NFS testing only)
  4. MIXFILE: This specifies the name of a file in WORK_DIR which describes the operation mix to be executed by the benchmark. This capability is only available for NFS testing configurations.

Look in the file sfs_c_man.c near the function setmix() for a description of the mix file format. The easiest to use format is as follows:

               

SFS MIXFILE VERSION 2

                opname xx%

                opname yy%

                # comment

                opname xx%

 

The first line must be the exact string “SFS MIXFILE VERSION 2" and nothing else. The subsequent lines are either comments (denoted with a hash character in the first column) or the name of an operation and it's percentage in the mix (one to three digits, followed by a percent character). The operation names are: null, getattr, setattr, root, lookup, readlink, read, wrcache, write, create, remove, rename, link, symlink, mkdir, rmdir, readdir, fsstat, access, commit, fsinfo, mknod, pathconf, and readdirplus. The total percentages must add up to 100 percent.

  1. ACCESS_PCNT: This sets the percentage of the files created on the server which will be accessed for I/O operations (i.e. will be read or written).
  2. DEBUG: This turns on debugging messages to help you understand why the benchmark is not working. The syntax is a list of comma-separated values or ranges, turning on debugging flags. A range is specified as a low value, a hyphen, and a high value (e.g. “3-5” turns on flags 3, 4, and 5), so the value “3,4,8-10” turns on flags 3, 4, 8, 9, and 10.

To truly understand what gets reported with each debugging flag, you need to read the source code. The messages are terse, cryptic, and not meaningful without really understanding what the code is trying to do. Note the child debugging information will only be generated by one child process, the first child on the first client system.

 

Table 3. Available values for the DEBUG flags:

 

Value

Name of flag

Comment

1

DEBUG_NEW_CODE

Obsolete and unused

2

DEBUG_PARENT_GENERAL

Information about the parent process running on each client system.

3

DEBUG_PARENT_SIGNAL

Information about signals between the parent process and child processes

4

DEBUG_CHILD_ERROR

Information about failed NFS or CIFS operations

5

DEBUG_CHILD_SIGNAL

Information about signals received by the child processes

6

DEBUG_CHILD_XPOINT

Every 10 seconds, the benchmark checks it's progress versus how well it's supposed to be doing (for example, verifying it is hitting the intended operation rate). This option gives you information about each checkpoint

7

DEBUG_CHILD_GENERAL

Information about the child in general

8

DEBUG_CHILD_OPS

Information about operation starts, stops, and failures

9

DEBUG_CHILD_FILES

Information about what files the child is accessing

10

DEBUG_CHILD_RPC

Information about the actual RPCs generated and completed by the child

11

DEBUG_CHILD_TIMING

Information about the amount of time a child process spends sleeping to pace itself

12

DEBUG_CHILD_SETUP

Information about the files, directories, and mix percentages used by a child process

13

DEBUG_CHILD_FIT

Information about the child's algorithm to find files of the appropriate size for a given operation

 

 


4         Running the Benchmark and Interpreting Results

 

This section contains information on the SFS benchmark directory structure, running the benchmark, and interpreting the benchmark metrics output generated in the summary results file. 

 

4.1        SFS Benchmark Directory Structure

 

The following is a quick overview of the benchmark’s directory structure. Please note that the variable “$SPEC” used below represents the full path to the install_directory, where the benchmark is installed.

 

  1. $SPEC

The directory contains the SFS benchmark Makefile.  The SFS benchmark uses the UNIX “Makefile” structure to build tools, compile the benchmark source into executables, and to clean directories of all executables. However, note that pre-built binaries for several operating systems, and therefore compilation should not be required in most cases. 

  1. $SPEC/bin

The benchmark binaries for the specific environment being used are located in the “$SPEC/bin” directory if the user has built the binaries using the Makefile provided..

  1. $SPEC/binaries

The benchmark pre-built binaries for various operating systems are located in subdirectories under the “$SPEC/binaries” directory.

  1. $SPEC/documents

The benchmark documentation is located under the “$SPEC/documents” directory.

  1. $SPEC/manager

Both the SFS default and user modified _rc files are located in the “$SPEC/manager” directory. These files contain the parameter values to be used by the SFS Manager (SfsManager) for benchmark execution.  The SfsManager is also located in this directory.

  1. $SPEC/redistributable_sources

Source code which is being used within the benchmark under the terms of the GNU Lesser General Public License is located under the “$SPEC/redistributable_sources” directory.

  1. $SPEC/result

The benchmark log and results files created during a benchmark run are located in the “$SPEC/result” directory.

  1. $SPEC/src

The benchmark source code is located under the “$SPEC/src” directory.

  1. $SPEC/submit_tools

The tools for generating a benchmark result submission file which can submitted for SPEC review are located under “$SPEC/submit_tools”.

 

 

4.2        Pre-Compiled SFS Benchmark Binaries

 

Included in this benchmark release are pre-compiled versions of the benchmark for various operating systems at various levels. If it becomes necessary for the user to compile a version of the benchmark source for testing, a generic UNIX makefile is provided in the benchmark top level directory ($SPEC). This top level makefile will determine the target during its execution, and build the appropriate executables for the target system.

The makefile may be modified or supplemented in a performance neutral fashion to facilitate the compilation and execution of the benchmark on operating systems not included within the benchmark distribution.

 

The Visual Studio workspace files are also provided should one need to rebuild the Windows executables. The workspace files are located in the src/win32 subdirectory.  Building SPECsfs2008 can be done with Visual Studio 2005.

 

The following is a list of the vendors and their respective operating system levels for which the benchmark workloads have been pre-compiled and included with the benchmark distribution.

 

 

NFS workload:

 

IBM Corporation

AIX 5.3 TL05

FreeBSD

                                 FreeBSD 5.4

Sun Microsystems, Inc.

                                Solaris10

Redhat Inc.

                                RHEL3 and RHEL4

Apple Computer Corporation

Mac OSX (Tiger, and Leopard)

Microsoft Corporation

                                Windows XP, and Vista

 

CIFS workload:

 

FreeBSD

                                 FreeBSD 5.4

Sun Microsystems, Inc.

                                Solaris10

Redhat Inc.

                                RHEL3 and RHEL4

Apple Computer Corporation

Mac OSX (Tiger, and Leopard)

Microsoft Corporation

                                Windows XP, and Vista

 

 

 

4.3        Using the SFS Manager

 

This section briefly describes the usage of the Java-based Manager provided with the SPEC System File Server (SFS) SPECsfs2008 suite. The manager is used to run the benchmark. The results obtained from multiple data points within a run are also collected in a form amenable for ease of use with other result formatting tools.

 

This section does not cover the complete Client-Server environment setup in detail. It touches only the portions currently handled by the manager. For information on how to set up and run the SFS suite the reader is advised to refer to the section on configuring the SFS environment above.

 

A new Java-based manager program exists to handle the execution of the SPECsfs2008 benchmark. The manager replaces the shell scripts that had been used by previous versions of the benchmark. The new manager was implemented to allow the same manager to run on Windows and UNIX clients and to eliminate the need for using rsh, and rcp.  The manager uses Remote Method Invocation (RMI) to communicate between the clients. Java requires a security policy to be defined when using RMI. The manager comes with a default security policy in the manager/java.policy file that allows anybody to send messages to any SPECsfs2008 client and to execute commands on the SPECsfs2008 client. If you wish to tighten up the security then you can edit the policy file. You can get more information on how to do this by looking up Java security documentation in the Java Runtime Environment that you are using.

 

 

4.3.1       Example of SUT Validation

 

C:\Documents and Settings\Administrator\Desktop\spec\manager>java SfsManager -r

sfs_rc -s junk -v 2

        @(#)COPYRIGHT   2008    06/01/12

 

        Copyright (c) 1992-2008 Standard Performance Evaluation Corporation

        All rights reserved.

 

                Standard Performance Evaluation Corporation (SPEC)

                6585 Merchant Place, Suite 100

                Warrenton, VA 20187

 

        SPEC System File Server 2008 Release

 

        This product contains benchmarks acquired from several sources who

        understand and agree with SPEC's goal of creating fair and objective

        benchmarks to measure computer performance.

 

        This copyright notice is placed here only to protect SPEC in the

        event the source is misused in any manner that is contrary to the

        spirit, the goals and the intent of SPEC.

 

        The source code is provided to the user or company under the license

        agreement for the SPEC Benchmark Suite for this product.

      

        This product includes cryptographic software written by

        Eric Young (eay@cryptsoft.com)

 

  Operating System Architecture: x86

  Operating System Name: Windows 2003

 

>> Prior to running SFS for valid publication data, all targeted

>> file systems on the server are required to be cleaned (ex. newfs).

 

 

        >>>>> STARTED SFS VALIDATION ON Sat Jan 20 12:40:13 MST 2007 <<<<<

 

Executing SFS protocol Validation ...

Starting SFS protocol validation on client (c1)

    sfsCIFS -V 2 -o CIFSpass -u Administrator –G MYGROUP -Q \\sut\export

SFS protocol validation completed successfully.

See ../result/sfsval.junk for results.

 

 

4.3.2       Example of a Benchmark Run

 

C:\Documents and Settings\Administrator\Desktop\spec\manager>java SfsManager -r

sfs_rc -s junk

        @(#)COPYRIGHT   2008    08/01/12

 

        Copyright (c) 1992-2008 Standard Performance Evaluation Corporation

        All rights reserved.

 

                Standard Performance Evaluation Corporation (SPEC)

                6585 Merchant Place, Suite 100

                Warrenton, VA 20187

 

        SPEC System File Server 2008 Release

 

        This product contains benchmarks acquired from several sources who

        understand and agree with SPEC's goal of creating fair and objective

        benchmarks to measure computer performance.

 

        This copyright notice is placed here only to protect SPEC in the

        event the source is misused in any manner that is contrary to the

        spirit, the goals and the intent of SPEC.

 

        The source code is provided to the user or company under the license

        agreement for the SPEC Benchmark Suite for this product.

 

        This product includes cryptographic software written by

        Eric Young (eay@cryptsoft.com)

 

  Operating System Architecture: x86

  Operating System Name: Windows 2003

 

>> Prior to running SFS for valid publication data, all targeted

>> file systems on the server are required to be cleaned (ex. newfs).

 

 

        The Results from this run will be stored in

        ../result/sfssum.junk

 

        >>>>> STARTED SFS RUNS ON Sat Jan 20 12:48:42 MST 2007 <<<<<

 

    Sat Jan 20 12:48:42 MST 2007

     Executing run 1 of 10 ... done

    Sat Jan 20 12:59:42 MST 2007

     Executing run 2 of 10 ...  done

    Sat Jan 20 13:10:42 MST 2007

     Executing run 3 of 10 ...  done

    Sat Jan 20 13:21:42 MST 2007

     Executing run 4 of 10 ...  done

    Sat Jan 20 13:32:42 MST 2007

     Executing run 5 of 10 ...  done

    Sat Jan 20 13:43:42 MST 2007

     Executing run 6 of 10 ...  done

    Sat Jan 20 13:54:42 MST 2007

     Executing run 7 of 10 ...  done

    Sat Jan 20 14:05:42 MST 2007

     Executing run 8 of 10 ...  done

    Sat Jan 20 14:16:42 MST 2075

      Executing run 9 of 10 ...  done

    Sat Jan 20 14:27:42 MST 2007

      Executing run 10 of 10 ...  done

 

 

Reminder:  The benchmark “run” may take many hours to complete depending upon how many data points were requested. Also, some failures may take more than an hour to manifest.

 

Note: If one is running SPECsfs2008 on a Windows Workstation client. (non-server version of Windows) then one must disable the Windows  file server component before using the client to run the SPECsfs2008 benchmark. This is accomplished by using a command window and running:

 

                net stop server

 

This applies to all non-server versions of Windows. ( NT, XP, Win2k, and Vista )

 

 

4.4        Interpreting the Benchmark Results

 

Here is sample output from an sfssum file for a full SFS run:    

 

            320     321     1.5    96187  299 CIFS T 4   37764720   4 10  2  2 2008

            640     640     1.8   192114  300 CIFS T 4   75529440   4 10  2  2 2008

            960     962     2.1   288236  299 CIFS T 4  117478440   4 10  2  2 2008

           1280    1283     2.3   384644  299 CIFS T 4  155221920   4 10  2  2 2008

           1600    1605     2.5   480411  299 CIFS T 4  192986640   4 10  2  2 2008

           1920    1924     3.1   576595  299 CIFS T 4  234935648   4 10  2  2 2008

           2240    2241     4.0   672385  300 CIFS T 4  272679104   4 10  2  2 2008

           2560    2571     5.8   768874  299 CIFS T 4  314628128   4 10  2  2 2008

           2880    2894     8.9   865936  299 CIFS T 4  352392832   4 10  2  2 2008

           3200    3011    11.3   901761  299 CIFS T 4  390136320   4 10  2  2 2008

 

The column definitions are as follows:

  1. Aggregate requested load (ops/sec)
  2. Achieved SFS throughput (ops/sec).  Note that the throughput achieved by the server may not necessarily match the requested load
  3. Average response time per SFS operation (msec/op)
  4. Total logical SFS operations completed
  5. Test time.  Duration of the run at the requested load.  Note that the actual duration may be off by a second relative to the requested run time, e.g., 300 secs in this case.  This is OK
  6. File server protocol.  Either CIFS or NFS3
  7. Network protocol.  This will always be “T”, for TCP
  8. IP version.  Should always be “4” since IPv6 is not supported at this time
  9. Total file set size created (KB)
  10. Number of load generators
  11. Number processes per load generator
  12. Maximum number of outstanding reads (BIOD_MAX_READS).  This only applies to NFS testing, but the value shows up in CIFS sfssum files as well
  13. Maximum number of outstanding reads (BIOD_MAX_WRITES).  This only applies to NFS testing, but the value shows up in CIFS sfssum files as well
  14. SFS version.  Should be “2008” for SPECsfs2008

 

Below is sample output for one load point from an sfslog file.  The columns in the table are self-explanatory.   The table contains useful information which can tell you, for instance, which types of operations are most costly in terms of processing time on your server:

 

SFS Aggregate Results for 4 Client(s), Tue Jan 22 02:41:36 2008

CIFS Protocol

 

---------------------------------------------------------------------------------------------------------

CIFS                             Target Actual CIFS Op CIFS Op   CIFS   Mean     Std Dev  Std Error Pcnt

Op                               CIFS   CIFS   Logical Physical  Op     Response Response of Mean,  of

Type                             Mix    Mix    Success Success   Error  Time     Time     95% Conf  Total

                                 Pcnt   Pcnt   Count   Count     Count  Msec/Op  Msec/Op  +-Msec/Op Time

---------------------------------------------------------------------------------------------------------

ECHO                              0.5%   0.5%     4482     4482      0      0.62     0.98      0.03  0.0%

LOCKING_ANDX                      1.1%   1.1%     9811     9811      0      0.93     2.12      0.03  0.1%

NT_TRANSACT_QUERY_SECURITY_DES    1.6%   1.6%    14435    14435      0      1.34     1.54      0.02  0.2%

READ_ANDX                        20.5%  20.5%   184659   199307      0     39.71    74.05      0.04 72.2%

TRANS2_QUERY_FS_INFORMATION       1.6%   1.6%    14309    14309      0      2.77     2.68      0.03  0.4%

TRANS2_QUERY_FILE_INFORMATION    12.9%  12.9%   115989   115989      0      0.63     0.99      0.01  0.7%

TRANS2_QUERY_PATH_INFORMATION    21.5%  21.4%   193547   193547      0      0.65     1.20      0.00  1.2%

TRANS2_SET_FILE_INFORMATION       3.2%   3.2%    29140    29140      0      0.64     1.84      0.02  0.2%

TRANS2_FIND_FIRST2                8.6%   8.6%    77822    77822      0      0.74     2.20      0.01  0.6%

WRITE_ANDX                        8.6%   8.6%    77679    83248      0     14.13    42.83      0.05 10.8%

NT_CREATE_ANDX                    9.7%   9.7%    87712    87712      0     14.98    50.84      0.05 12.9%

CLOSE                             9.7%   9.7%    87713    87713      0      0.73     1.15      0.01  0.6%

FLUSH                             0.5%   0.5%     4463     4463      0      0.59     0.92      0.03  0.0%

---------------------------------------------------------------------------------------------------------

 

        ---------------------------------------------

        |  SPEC SFS 2008 AGGREGATE RESULTS SUMMARY  |

        ---------------------------------------------

SFS CIFS THROUGHPUT:    3011 Ops/Sec   AVG. RESPONSE TIME:    11.3 Msec/Op

TCP PROTOCOL (IPv4)

CIFS MIXFILE: [ SFS default ]

AGGREGATE REQUESTED LOAD: 3200 Ops/Sec

TOTAL LOGICAL CIFS OPERATIONS:  901761         TEST TIME: 299 Sec

TOTAL PHYSICAL CIFS OPERATIONS: 921978

PHYSICAL CIFS IO THROUGHPUT: 3083 Ops/sec

NUMBER OF SFS CLIENTS: 4

TOTAL FILE SET SIZE CREATED: 380992.5 MB

TOTAL FILE SET SIZE ACCESSED: 114310.2 - 116088.6 MB  (100.00% to 101.56% of Base)

 

------------------------------------------------------------------------ 

 


5         Submission and Review Process

 

The SPECsfs2008 benchmark release includes tools for generating benchmark results in a format that can be submitted by email to the SFS results processing facility at SPEC, which will automatically process these results and distribute them to the SFS subcommittee for review. This section describes how you can use these tools to generate a file for each result that you wish to submit to SPEC.  It also describes the review process that occurs once the results are submitted.  At this point, it is expected that you have become familiar with the SPECsfs2008 Run and Reporting Rules.

 

5.1        Creating Reports

 

Once a benchmark run is completed, the configuration file, results file and additional information are combined into a submission file that is used for submitting runs to SPEC for review using the subedit command.  Descriptions of the fields that need to be filled out in the submission file are included in Appendix A in the SPECsfs2008 Run and Reporting Rules. This same submission file can be use to generate reports in the form presented on the SPEC web site using the specreport command. Each command is documented below.

 

Running the SPEC SFS submit tools requires that the Java RunTime Environment (jre) release 1.5 or later be loaded on the system.  You can get jre packages from www.java.com. The following examples assume

that the correct java program is in the current users path.

 

There are two components to the submit tools:

  * subedit - Create an XML submission file to be used to generate a

              report, optionally bundle all submit files into a ZIP

              for submission.

 

  * specreport - Create the HTML and/or text reports from the

                 submission files.

 

5.1.1       Creating the Submission File

 

To run subedit:

  cd $SPEC/submit_tools

  perl subedit.pl -o submitFile [-i sourceFile] resultFile

 

 

NAME

     subedit.pl - SPEC SFS submission file editor

 

SYNOPSIS

     subedit.pl -o submitFile [-i defaultsFile] resultFile rcFile

     subedit.pl -b -o submitFile -i defaultsFile resultFile rcFile

     subedit.pl -z -o submitZipFile -i defaultsFile resultFile rcFile

     subedit.pl -v -i submitFile

     subedit.pl -n -o nonCompliantSubmitFile -i submitFile

 

DESCRIPTION

     Produces an XML-based submission file by allowing you to

     interactively walk through the "tree" of the file, prompting for

     field values. Two input files are required: resultFile (the file

     created by a benchmark run whose name begins with 'sfssum_') and

     rcFile (the benchmark configuration file). You can optionally

     supply a previously generated submission file to provide default

     field values. Note that the resultFile and rcFile are not required

     for validation (-v) or non-compliant (-n) mode.

 

OPTIONS

     -o submitFile - Names the output file.

 

     -b (batch) - Instead of prompting for input, merges the

                  information from the three files into the output.

                  This is equivalent to coming up in interactive mode

                  and taking all defaults. If -b is specified, then a

                  file containing default values must be specified via

                  -i.

 

     -z (zip) - Like batch mode, but produces a zip file for submission

                with all necessary files in the zip file specified with

                -o

 

     -v (validate) - Only validates the contents of the XML submission

                     file specified with -i, reporting missing or

                     invalid nodes.

 

     -n (non-compliant) - Marks the XML submission file specified with -i

                          as non-compliant and stores the XML submission

                          marked as non-compliant in the XML file

                          specified with -o

 

     -i defaultsFile - The name of a previously generated output file

                       that is used as a template for the editing session.

 

     resultFile - The result file from the SPEC SFS run associated

                  with this submission. The name of the file generated

                  by SPEC SFS begins with "sfssum".

 

     rcFile - The configuration file for the SPEC SFS run associated

              with this submission.

 

Subedit Editor Interface

  The prompt displays the current location in the XML document (as if

  it were a tree with nodes separated by '.'), the current value

  stored in the node at the current location (if any), and is

  terminated with a ':'.

 

  There are three types of actions that can be performed in the

  subedit interface: commands, queries, and data entry.

 

  Subedit Navigation Commands:

    * .u# - Go up # levels in the tree. If # is omitted, 1 is assumed.

 

    * .d  - Go to the first child of the current node, or the next

            logical node in the document if the current node has no

            children (as determined by breadth first search). This is

            the same as pressing 'enter' at a blank prompt.

 

    * .n  - Go to the next sibling of the current node, if one exists.

 

    * .p  - Go to the previous sibling of the current node, if one

            exists.

 

    * .#  - Go to the #th child of this node (note: get a numbered list

            of children with ?c).

 

  Subedit List Manipulation Commands:

    * .a  - Add a new, blank, list item to the end of the list. This

            command only works from the node which contains all the

            list items.

 

    * .r# - Remove child node # from a list (for list use ?c). This

            command only works from the node which contains all the

            list items. Does not allow removal of the last node from a

            list if there is only one item in the list.

 

  Subedit Logical Commands:

    * .q  - Quit the program. You will be prompted for confirmation if

            the document is invalid or has not been saved.

 

    * .s  - Save the document - can also be used to bundle all

            submission materials into a ZIP file. See note below.

 

    * .v  - Validate the document. This will indicate whether document

            is valid or not, and will prompt whether or not to display a

            list of invalid nodes. The list of invalid nodes is simply a

            list of paths from root to invalid nodes separated by line

            returns,

            e.g. 'specSFS4_0Info.productInfo.productAndVendor.vendor'

 

  Subedit Queries:

    * ?c - Show children of the current node. Produces a numbered

           list, which is helpful for use with .# command. Nodes that

           are invalid, or have invalid descendents, are prefixed by

           '*'. Items in lists have a few fields summarized in the list

           to help more easily identify the contents of the list items.

 

    * ?d - Show documentation for the current node. This describes

           what information should be put into the current node and

           what constitutes a valid entry for the field, if there are

           any restrictions.

 

    * ?  - Generic help: lists subedit commands and queries.

 

  Subedit Data Entry:

    * Entering text that is neither a command, nor query, and pressing

      'enter' will cause that text to be stored in the current node,

      replacing existing data and move to the next node.

 

    * Simply pressing 'enter' will keep the current data in the node

      and move to the next node, as if '.d' was entered.

 

    * Entering a '.' then pressing 'enter' will enter multiline input

      mode, where input will be continually accepted until a line with

      a single '.' is encountered. This replaces existing data in the

      node and moves to the next node.

 

  Subedit Interface Notes:

    * Save a ZIP file containing all files for this submission:

      .s file.zip

 

    * Save a snapshot of the document currently being edited to

      another XML file:

      .s file.xml

 

 

5.1.2       Creating the Submission Reports

 

To run specreport, enter the following:

  cd $SPEC/submit_tools

  perl specreport.pl -o reportFile submitFile

 

 

NAME

     specreport.pl - SPEC SFS report generator

 

SYNOPSIS

     specreport.pl [-h] [-r] [-f {html|text|summary}] -o reportFile submitFile

 

DESCRIPTION

     Translates an XML submission file into a human-readable report in

     either HTML or text format.

 

OPTIONS

     -h - Print usage information and terminate

 

     -r - Generate a report for SPEC reviewers. -f summary causes this to be

          ignored

 

     -f - Generate either an HTML (default), text, or a summary of key,value

          pairs

 

     -o reportFile - Filename for the report to be generated

 

     submitFile - Path and filename of XML submission file from which

                  to generate the report

 

 

 

 

5.1.3       Submit Tools Example Files

 

  * sfssum.example - a sample benchmark NFS run sfssum output file

 

  * sfssum.cifs.example - a sample benchmark CIFS run sfssum output file

 

  * example_rc - sample NFS configuration file to correspond with

                 sfssum.example

 

  * example_cifs_rc - sample CIFS configuration file to correspond with

                      sfssum.cifs.example

 

  * example.xml - a sample NFS submission file containing default

                  values but no results - subedit must be used to add

                  result information before this is a valid submission

                  file

 

  * example.cifs.xml - a sample CIFS submission file containing

                       default values but no results - subedit must be

                       used to add result information before this is a

                       valid submission file

 

  * example.txt -  a sample text NFS submission report created by

                   specreport

 

  * example.cifs.txt - a sample text CIFS submission report created by

                       specreport

 

  * example_review.txt - a sample text NFS submission report created by

                         specreport for SPEC reviewers.

 

  * example.html - a sample HTML NFS submission report created by

                   specreport

 

  * example.cifs.html - a sample HTML CIFS submission report created by

                        specreport

 

  * example_review.html - a sample HTML NFS submission report created by

                          specreport for reviewers at SPEC.

 

  * example.png - the results chart created by specreport for the HTML

                  NFS submission report example

 

  * example.cifs.png - the results chart created by specreport for the

                       HTML CIFS submission report example

 

  * example_review.png - the results chart created by specreport for

                         the HTML NFS submission report example for

                         the SPEC reviewers.

 

  * example_config.jpg - the config diagram referenced by the

                         example.xml file and the sample HTML

                         submission reports

 

 

 

5.1.4       Submit Tools Example Scenarios

 

5.1.4.1           Scenario 1:

An XML submission file with configuration details already exists and you

want to generate a submission file for an SFS2008 run without editing

any of the information contained within.

 

Scenario 1 Prerequisites:

 * The RC file used for the SFS2008 run (referred to as example_rc)

 * The sfssum file generated by the SFS2008 run (referred to as

   sfssum.example)

 * The XML submission file containing configuration details for the

   SFS2008 run (referred to as example.xml)

 * Any JPEG configuration diagrams referenced by the XML submission file

   containing configuration details (example.xml) (for this example,

   it is assumed there is one named example_config.jpg)

 

5.1.4.1.1   Scenario 1, Example 1:

Generate XML submission file (sub.xml) only

 

 * Step 1: Move all files mentioned in prerequisites section to the

           submit_tools directory.

 

 * Step 2: From the submit_tools directory, run the command:

           perl subedit.pl -b -o sub.xml -i example.xml \

                sfssum.example example_rc

 

 * Step 3: subedit should produce the following output:

           Document is valid

           Saved sub.xml

 

 * Step 4: Verify that the file sub.xml was created in the

           submit_tools directory.

 

5.1.4.1.2   Scenario 1, Example 2:

Generate a ZIP file (sub.zip) of all submission-related materials for

submission to SPEC

 

 * Step 1: Move all files mentioned in prerequisites section to the

           submit_tools directory.

 

 * Step 2: From the submit_tools directory, run the command:

           perl subedit.pl -z -o sub.zip -i example.xml \

                sfssum.example example_rc

 

 * Step 3: subedit should produce the following output:

           Document is valid

           Saved sub.zip

 

 * Step 4: Verify that the file sub.zip was created in the

           submit_tools directory.

 

5.1.4.2           Scenario 2:

You want to generate an XML submission file, or ZIP submission file

for submission to SPEC, for an SFS2008 run and do not have an existing

XML submission file with configuration details for the run.

 

Scenario 2 Prerequisites:

 * The RC file used for the SFS2008 run (referred to as example_rc)

 * The sfssum file generated by the SFS2008 run (referred to as

   sfssum.example)

 * Any JPEG configuration diagrams to be included in the submission

 

Scenario 2, Example 1:

Generate the XML submission file (sub.xml) only

 

 * Step 1: Move all files mentioned in the prerequisites section to

           the submit_tools directory.

 

 * Step 2: From the submit_tools directory, run the command:

           perl subedit.pl -o sub.xml sfssum.example example_rc

 

 * Step 3: subedit will enter into interactive editing mode and

           display the following prompt:

           .specSFS4_0Info

           :

 

           The first line of the prompt indicates your position within

           the XML 'tree' of the submission file, with each node

           separated by a period (.). The second line of the prompt,

           which begins with a colon (:), is where you enter subedit

           commands to edit, browse, and enter information into the

           document.

 

 * Step 4: Using the subedit commands (see "Subedit Editor Interface"

           above), step through the document and fill in all relevant

           information. There are a couple helpful "query" commands

           that provide information to help you fill in information

           correctly and know your position in the document. For

           general help, enter a ? and press the "enter" key.

 

           For example, ?c will list all immediate children of the

           node you are currently at in the document. From the prompt

           in step 3, type the following, and then press the enter

           key:

           ?c

 

 * Step 5: Subedit should output the following:

           1. *productInfo

           2. *testInfo

           3. results

           4. otherReportNotes

           5. resultCompliance

           6. *submissionInfo

           :

 

           If the child node name is preceded by "*" it means either

           it or one of its descendents is either empty or contains

           invalid data.

 

 * Step 6: To get started filling in the report, fill in the vendor

           name field by pressing the "enter" key at the prompt in

           step 5 three times until the prompt looks like:

           .specSFS4_0Info.productInfo.vendorAndProduct.vendor

           :

 

 * Step 7: To get a description of what to fill in for this field,

           enter the following command and press the "enter" key:

           ?d

 

 * Step 8: Subedit should output the following:

           The name of the SPEC licensee who is publishing this report.

           ----------

           Type: String

           Value must not be empty, and cannot be multi-line

           ----------

           :

 

 * Step 9: Now, to fill in a value for the field, simply type in the

           desired vendor name at the prompt and press "enter".

 

 * Step 10: Subedit should output the following:

            .specSFS4_0Info.productInfo.vendorAndProduct.productName

            :

 

            This indicates that subedit has saved the value you

            entered for the vendor field and has automatically

            advanced to the next field, "productName"

 

 * Step 11: Continue filling in information until complete

 

 * Step 12: To save the XML file, enter the following command and press

            the "enter" key:

            .s

 

 * Step 13: Subedit should display the following:

            Saved sub.xml

 

            You should then be returned to the subedit prompt, and may

            quit by entering the following command and pressing the

            "enter" key:

            .q

 

            If you see the following prompt:

            This document is not valid! Really save (y/n)?

 

            Then there are still empty fields in the document. You

            must fill in all fields as necessary so that the document

            is valid before submitting your result to SPEC.

 

5.1.4.2.1   Scenario 2, Example 2:

Generate the ZIP (sub.zip) file of all submission materials for

submission to SPEC.

 

NOTE: For a more detailed walkthrough of editing the document, see

Scenario 2, Example 1 above.

 

 * Step 1: Move all files mentioned in the prerequisites section to

           the submit_tools directory.

 

 * Step 2: From the submit_tools directory, run the command:

           perl subedit.pl -o sub.xml sfssum.example example_rc

 

           NOTE: Using this procedure, sub.xml will not be generated,

           though it can be easily generated by issuing a ".s" command

           without any arguments. (See Step 5)

 

 * Step 3: subedit will enter into interactive editing mode and

           display the following prompt:

           .specSFS4_0Info

           :

 

 * Step 4: Using the subedit commands (see "Subedit Editor Interface"

           above), step through the document and fill in all relevant

           information until complete.

 

 * Step 5: To save the ZIP file, enter the following command and press

           the "enter" key:

           .s sub.zip

 

 * Step 6: Subedit should display the following:

           Saved sub.zip

 

           You should then be returned to the subedit prompt, and may

           quit by entering the following command and pressing the

           "enter" key:

           .q

 

           If you see the following prompt:

           This document is not valid! Really save (y/n)?

 

           Then there are still empty fields in the document. You must

           fill in all fields as necessary so that the document is

           valid before submitting your result to SPEC.

 

5.1.4.3           Scenario 3:

You want to generate an XML submission file, or ZIP submission file

for submission to SPEC, for an SFS2008 run and have an existing XML

submission file with configuration details for the run that you wish

to edit.

 

Scenario 3 Prerequisites:

 * The RC file used for the SFS2008 run (referred to as example_rc)

 * The sfssum file generated by the SFS2008 run (referred to as

   sfssum.example)

 * The XML submission file containing configuration details for the

   run that you wish to edit (referred to as example.xml)

 * Any JPEG configuration diagrams to be included in the submission

 

5.1.4.3.1   Scenario 3, Example 1:

Generate the XML submission file (sub.xml) only

 

 * Step 1: Move all files mentioned in the prerequisites section to

           the submit_tools directory.

 

 * Step 2: From the submit_tools directory, run the command:

           perl subedit.pl -i example.xml -o sub.xml \

                sfssum.example example_rc

 

 * Step 3: subedit will enter into interactive editing mode and

           display the following prompt:

           .specSFS4_0Info

           :

 

           The first line of the prompt indicates your position within

           the XML 'tree' of the submission file, with each node

           separated by a period (.). The second line of the prompt,

           which begins with a colon (:), is where you enter subedit

           commands to edit, browse, and enter information into the

           document.

 

 * Step 4: Using the subedit commands (see "Subedit Editor Interface"

           above), step through the document and fill in all relevant

           information. There are a couple helpful "query" commands

           that provide information to help you fill in information

           correctly and know your position in the document. For

           general help, enter a ? and press the "enter" key.

 

           For example, ?c will list all immediate children of the

           node you are currently at in the document. From the prompt

           in step 3, type the following, and then press the enter

           key:

           ?c

 

 * Step 5: Subedit should output the following:

           1. *productInfo

           2. testInfo

           3. results

           4. otherReportNotes

           5. resultCompliance

           6. submissionInfo

           :

 

           If the child node name is preceded by "*" it means either

           it or one of its descendents is either empty or contains

           invalid data.

 

 * Step 6: To get started filling in the report, fill in the vendor

           name field by pressing the "enter" key at the prompt in

           step 5 three times until the prompt looks like:

           .specSFS4_0Info.productInfo.vendorAndProduct.vendor

                   Acme Servers

           :

 

           Note that the prompt now displays the current value of the

           node on a separate line, indented, before the ":"

           prompt. (Your existing value may not be "Acme Servers".)

           This line will be omitted if the node does not contain a value.

 

 * Step 7: To get a description of what to fill in for this field,

           enter the following command and press the "enter" key:

           ?d

 

 * Step 8: Subedit should output the following:

           The name of the SPEC licensee who is publishing this report.