| 
   
    
      Results 
    
     SPECapc 
    
     Download/Order 
    
     Resources 
    
  
  
   | 
      
 
    
    
    
    The Application Performance Characterization Project
    Committee Rules   
Version 2.10 
Last Updated: 09/13/2007 
    - Overview 
        
            - General Philosophy 
                
                    - All rules declared in "The
                            Graphics and Workstation Performance Group (SPEC/GWPG):
                            Rules For Project Groups" document (known
                            hereafter as the GWPG Project Groups Ruleset)
                            apply, unless specifically overruled by a rule
                            in this document. 
 
                 
             
            - General Philosophy 
                
                    - Within SPEC's Graphics and Workstation Performance
                        Group (SPEC/GWPG) there was a strong belief that
                        it is important to benchmark graphics performance
                        based on actual applications. Application-level benchmarks
                        exist, but they are not standardized and they do
                        not cover a wide range of application areas. Thus,
                        the Application Performance Characterization (SPECapcSM)  Project
                        was created within the GWPG to create a broad-ranging
                        set of standardized benchmarks for graphics-intensive
                        applications. 
 
                    - The SPECapc seeks to develop benchmarks for generating
                        accurate application-level graphics performance measures
                        in an open, accessible and well-publicized manner. 
 
                    - The SPECapc wishes to contribute to the coherence
                        of the field of application performance measurement
                        and evaluation so that vendors will be better able
                        to present well-defined performance measures and
                        customers will be better able to compare and evaluate
                        vendors products and environments. 
 
                    - The SPECapc will provide formal beta benchmarks to
                        members and final benchmark releases to the public
                        in a timely fashion. 
 
                    - Hardware and software used to run the SPECapc benchmarks
                        must provide a suitable environment for running typical
                        (not just benchmark) workloads for the applications
                        in question. 
 
                    - SPECapc reserves the right to adapt its benchmarks
                        as it deems necessary to preserve its goal of fair
                        and useful benchmarking (e.g. remove benchmark, modify
                        benchmark code or data, etc). If a change is made
                        to the suite, SPECapc will notify the appropriate
                        parties (i.e. SPECapc members and users of the benchmark)
                        and SPECapc will re-designate the benchmark by changing
                        its name and/or version. In the case that a benchmark
                        is removed in whole or in part, SPECapc reserves
                        the right to republish in summary form "adapted" results
                        for previously published systems, converted to the
                        new metric. In the case of other changes, such a
                        republication may necessitate re-testing and may
                        require support from the original test sponsor. 
 
                 
             
            - Overview of Optimizations 
                
                    - SPECapc is aware of the importance of optimizations
                        in producing the best system performance. SPECapc
                        is also aware that it is sometimes hard to draw an
                        exact line between legitimate optimizations that
                        happen to benefit SPECapc benchmarks and optimizations
                        that specifically target SPECapc benchmarks. However,
                        with the list below, SPECapc wants to increase awareness
                        of implementers and end-users to issues of unwanted
                        benchmark-specific optimizations that would be incompatible
                        with SPECapc's goal of fair benchmarking. 
 
                    - To ensure that results are relevant to end-users,
                        SPECapc expects that the hardware and software implementations
                        used for running SPECapc benchmarks adhere to a set
                        of general rules for optimizations. 
 
                 
             
            - General Rules for Optimization 
                
                    - Optimizations must generate correct images and results
                        for the application under test, for both the benchmark
                        case and similar cases. Correct images and results
                        are those deemed by the majority of the SPECapc electorate,
                        potentially with input from the associated independent
                        software vendor (ISV) and/or end-users, to be sufficiently
                        adherent to the intent behind the application. 
 
                    - Optimizations must improve performance for a class
                        of workloads where the class of workloads must be
                        larger than a single SPECapc benchmark or SPECapc
                        benchmark suite. 
 
                    - For any given optimization a system should generate
                        correct images with and without said optimization.
                        An optimization should not reduce system stability. 
 
                    - The vendor encourages the implementation for general
                        use (not just for running a single SPECapc benchmark
                        or SPECapc benchmark suite). 
 
                    - The implementation is generally available, documented
                        and supported by the providing vendor. 
 
                    - In the case where it appears that the above guidelines
                        have not been followed, SPECapc may investigate such
                        a claim and request that the optimization in question
                        (e.g. one using SPECapc benchmark-specific pattern
                        matching) be removed and the results resubmitted.
                        Or, SPECapc may request that the vendor correct the
                        deficiency (e.g. make the optimization more general
                        purpose or correct problems with image generation)
                        before submitting results based on the optimization. 
 
                    - It is expected that system vendors would endorse
                        the general use of these optimizations by customers
                        who seek to achieve good application performance. 
 
                    - No pre-computed (e.g. driver-cached) images, geometric
                        data, or state may be substituted within an SPECapc
                        benchmark on the basis of detecting that said benchmark
                        is running (e.g. pattern matching of command stream
                        or recognition of benchmark's name). 
 
                 
             
         
     
 
    - Benchmarks 
        
            - Benchmark Acceptance 
                
                    - Benchmark components are defined as
                        
                            - specific revision of an application, 
 
                            - run rules,  scripts and associated data
                                sets. 
 
                         
                     
                    - New or modified benchmark components require a 2/3-majority
                        vote to be accepted for publication. Selection of
                        datecode versions of a specific revision of an application
                        is by majority vote. 
 
                    - A minimum 3-week review period is required for new
                        or significantly modified benchmark components. 
 
                    - At the end of the review period a vote will be called
                        to approve the proposed changes. 
 
                    - An amendment to a benchmark component during the
                        review period must be unanimously accepted. If not,
                        the review period shall be restarted. 
 
                 
             
            - Benchmark Code Versioning 
                
                    - Benchmarks use the following version coding: M.m
                        (e.g. SPECapcSM for Pro/ENGINEER 20.0
                        v1.1) M is the major release number and m is the
                        minor release number. 
 
                    - The major release number is only incremented when
                        large amounts of code are changed and the scripting
                        language is dramatically changed as a result -- backward
                        compatibility is highly unlikely when moving scripts
                        or data sets between major releases (e.g. running
                        v2 scripts on a v3 executable would almost certainly
                        fail). 
 
                    - The minor release number is bumped if some small
                        set of code is replaced or removed - but the standard,
                        unchanged scripts and data sets, as a whole, must
                        run on the new version (but perhaps with different
                        performance). 
 
                    - When there is a new major release of a benchmark,
                        submissions using the previous release will be accepted
                        for at least one submission cycle. When a superceded
                        benchmark is retired, the associated run-rules will
                        be archived in an "archived benchmark run-rules
                        document" posted on the Previous SPECapc Benchmarks
                        web-page. 
 
                 
             
         
     
 
    - Submission, Review and Publication
        
            - General Benchmark Run Rules 
                
                    - The system under test must correctly perform all
                        of the operations being requested by the application
                        during the benchmark. 
 
                    - No changes to any files associated with the benchmark
                        are permitted excepted as noted in the benchmark-specific
                        rules (section 5 of this document). 
 
                    - The entire display raster must be available for use
                        by the application being benchmarked. 
 
                    - It is not permissible to override the intended behavior
                        of the application through any means including, but
                        not limited to, registry settings or environment
                        variables. 
 
                    - No interaction is allowed with the system under test
                        during the benchmark, unless required by the benchmark. 
 
                    - The system under test can not skip frames during
                        the benchmark run. 
 
                    - It is not permissible to change the system configuration
                        during the running of a given benchmark. That is,
                        one can't power off the system, make some changes,
                        then power back on and run the rest of the benchmark. 
 
                    - Results submitted must be obtained using the scripts,
                        models, and application revisions which are specified
                        for that submission cycle by the SPECapc. 
 
                    -  The color depth used must be at least 24 bits (true
                        color), with at least 8 bits of red, 8 bits of green
                        and 8 bits of blue, unless otherwise specified. 
 
                    -  The display raster resolution must be at least 1280
                        pixels by 1024 pixels, unless otherwise specified. 
 
                    -  The monitor refresh rate must be at least 75Hz.
                        This requirement does not apply to digital flat panel
                        displays, unless otherwise specified. 
 
                    - The border width of the windows created during the
                        benchmark shall not exceed 10 pixels, unless otherwise
                        specified. 
 
                    - The monitor used in the benchmark must support the
                        stated resolution and refresh rate. 
 
                    - The benchmark must successfully obtain all requested
                        window sizes, with no reduction or clipping of any
                        benchmark-related windows.  Windows created
                        by the benchmark must not be obscured on the screen
                        by anything other than other elements created by
                        the benchmark. 
 
                    - Tests may be run with or without a desktop/window
                        manager if the application allows this, but must
                        be run on some native windowing system. 
 
                 
             
            - General Submission Content Rules 
                
                    - The submission upload file structures are defined
                        in the benchmark-specific section below. 
 
                 
             
            - Submission Process Rules 
                
                    - The submission file names are detailed below under
                        the benchmark-specific rules. 
 
                 
             
            - Review Period Rules 
                
                    - Reviewers will decide if the image quality and results
                        of the submission are sufficiently correct with respect
                        to the intent of the ISV to satisfy the intended
                        end-users' expectations. 
 
                 
             
         
     
 
    - SPECapc Benchmark Specific Rules and Procedures
        
            - Solid Edge V14
                
                    - The benchmark must be run using Solid Edge V14.00.00.70 
 
                    - The application can be run with Tools-Options-View
                        Graphics Display set to Graphics Card Driven, Software
                        Driven or Backing Store. The choice must be documented
                        in the notes portion of the results. 
 
                    - Application settings must not be changed from the
                        defaults set by the benchmark installation.  Settings
                        that must not be changed include, but are not limited
                        to:
                        
                            - Culling 
 
                            - Wireframe display in Move Part command 
 
                            - Arc smoothness (3) 
 
                         
                     
                    - The submission must contain the file result.txt that
                        is generated during the benchmark run. 
 
                    - The directory structure of the submission  must
                        be as follows:
 
                        .../company-name/system_1/SEV14/result.txt 
                        .../company-name/system_2/SEV14/result.txt 
                        etc...  
                    - The submission file must be named company_apc_SolidEdgeV14_vN.zip
                        where company is the member company or organization
                        name in lower case and vN is the file version
                        (e.g. hp_apc_SolidEdgeV14_v0.zip.) The initial submission
                        is v0. Resubmitted files must have the version number
                        incremented. 
 
                 
             
            - Solidworks 2005
                
                    - The benchmark must be run using Solidworks2005 service
                        pack 0. 
 
                    - The application window size must not be changed from
                        its initial size. 
 
                    - The submission must contain a results.txt file generated
                        by running the benchmark. Its contents are extracted
                        from the file apcresultsN.txt, generated by
                        benchmark GUI , where N is the number
                        of the benchmark test run. 
 
                    - The submission will be derived from the best composite
                        generated by the default 5 benchmark runs, as controlled
                        and reported by the benchmark GUI. 
 
                    - The appearance of the the application's Quick Tips
                        / Dynamic Help box, or other pop-ups that do not
                        dismiss themselves, will cause the benchmark run
                        to be invalid.  The benchmark FAQ has information
                        on how to prevent this. 
 
                    - The directory structure of the submission  must
                        be as follows:
 
                        .../company-name/system_1/sw2005/results.txt 
                        .../company-name/system_2/sw2005/results.txt 
                        etc...  
                    - The submission file must be named company_apc_sw2005_vN.zip
                        where company is the member company or
                        organization name in lower case and vN is
                        the file version (e.g. ibm_apc_sw2005_v0.zip.) The
                        initial submission is v0. Resubmitted files must
                        have the version number incremented. 
 
                 
             
            - 3dsmax7
                
                    -  The benchmark must be run using 3ds max
                        version 7, service pack 0 (i.e. unpatched). 
 
                    - The application windows will be visible as specified:
                        
                            -  The "3ds max 7" window
                                is maximized and fills the screen (The window
                                is not occluded by any other windows.) Note:
                                Task-bar auto-hide should be enabled. 
 
                         
                     
                    - Drivers and custom drivers must be configured (explicitly
                        or implicitly) to use:
                        
                            -  GL_TEXTURE_MIN_FILTER = GL_LINEAR_MIPMAP_LINEAR 
 
                            - GL_TEXTURE_MAG_FILTER = GL_LINEAR 
 
                            - Textures cannot be modified, i.e. reduced
                                in size or depth. 
 
                            - Wireframe objects must NOT be drawn using
                                triangle strips. 
 
                         
                     
                    - The following Viewport parameters must be checked
                        (i.e. enabled) in the Viewports Tab found in the
                        3ds max 7 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Backface cull on object creation 
 
                            - Mask viewport to safe region 
 
                            - Update background while playing 
 
                            - Display world axis 
 
                         
                     
                    - The following Viewport parameters must NOT be checked
                        in the 3ds max 7 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Attenuate lights 
 
                            - Filter environment backgrounds 
 
                         
                     
                    - The benchmark can be run using either OpenGL or Direct
                        3D. 
 
                    - For OpenGL
                            submissions, an application graphics driver or
                            plug-in may be used, but the images captured
                            from the benchmark must not differ from the supplied
                            3ds max OpenGL driver's images by more than 5%.
                            If the application driver or plug-in’s
                            images do not match, the submitter may apply
                            for a waiver based on documented errors in the
                            3ds max OpenGL driver. The availability and support
                            of an application driver or plug-in shall not
                            affect the submission's “single supplier” or “parts
                            built” status. 
 
                    - The submission must contain the files result.txt
                        and result.xls that are generated during the benchmark
                        run. 
 
                    - The directory structure of the submission  must
                        be as follows:
 
                        .../company-name/system_1/3dsmax7/result.txt  
                        .../company-name/system_1/3dsmax7/result.xls  
                        .../company-name/system_2/3dsmax7/result.txt 
                        etc...  
                    - The submission file must be named company_apc_3dsmax7_vN.zip
                        where company is the member company or organization
                        name in lower case and vN is the file version
                        (e.g. hp_apc_3dsmax7_v0.zip.) The initial submission
                        is v0. Resubmitted files must have the version number
                        incremented. 
 
                    - Submitters are not required to include pixel-comparison
                        images. 
 
                    - A committee member raising a challenge to a submission
                        can employ pixel-comparison images to support the
                        challenge. However, reference and grabbed images
                        must be generated on a system with identical GPU,
                        graphics and custom drivers and settings to the submitted
                        configuration being challenged. 
 
                 
             
            - Maya 6.5
                
                    -  The benchmark must be run using Maya 6.5 
 
                    - The benchmark script must be run with the command
                        'mayaTest(3)' where '3' is the number of runs. 
 
                    - The submission must contain the results.txt submission
                        file as well as the scoring spreadsheet used to calculate
                        the result. 
 
                    - The directory structure of the submission  must
                        be as follows:
 
                        .../company-name/system_1/maya65/results.txt  
                        .../company-name/system_1/maya65/MayaResults.xls 
                        etc...  
                    - The submission file must be named company_apc_maya65_vN.zip
                        where company is the member company or
                        organization name in lower case and vN is
                        the file version (e.g. ibm_apc_maya65_v0.zip.) The
                        initial submission is v0. Resubmitted files must
                        have the version number incremented. 
 
                 
             
            - Pro/ENGINEER Wildfire 2.0
                
                    - The benchmark must be run using
                            build “M160” of Pro/ENGINEER Wildfire
                            2.0. Submissions may be made with 32-bit or 64-bit
                            versions of M160. However, submissions made with
                            32-bit Pro/ENGINEER Wildfire 2.0 on Windows XP
                            64-bit Edition are prohibited. 
 
                    - The config.pro file must be
                            used as-is and may not be modified or overridden.  
 
                    - The script files “utilities\runbench.bat” or “utilities/runbench.csh” may
                            be modified as necessary to enable execution
                            of the benchmark on the system being tested.
                            If modified, the modified version must be included
                            in the benchmark submission. 
 
                    - The submission must contain
                            the “proe_result.txt” file (as generated
                            by the “proescore” program) as well
                            as the corresponding “trail.txt” file
                            generated from running the benchmark. Both of
                            these files may be found in the "results" directory
                            after a successful run of the benchmark. The
                            first section of the “proe_result.txt” file
                            must be edited to reflect the system configuration. 
 
                    - There must be no license-related
                            warnings in the “trail.txt” file. 
 
                    - The directory structure of
                            the submission must be as follows: 
 
                        .../company-name/system_1/proewildfire2/proe_result.txt 
                        .../company-name/system_1/proewildfire2/trail.txt
                        (may be compressed) 
                        .../company-name/system_1/proewildfire2/runbench.bat
                        (or runbench.csh, if modified, as required by rule
                        c above) 
                        .../company-name/system_2/proewildfire2/proe_result.txt 
                        .../company-name/system_2/proewildfire2/trail.txt
                        (may be compressed) 
                        .../company-name/system_2/proewildfire2/runbench.bat
                        (or runbench.csh, if modified, as required by rule
                        c above) 
                        etc...  
                    - The submission file must be
                            named company_apc_proewildfire2_vN.zip
                            or company_apc_proewildfire2_vN.tar.z
                            where company is the member company or
                            organization name in lower case and vN is
                            the file version (e.g. “sgi_apc_proewildfire2_v0.tar.z” and “intel_apc_proewildfire2_v0.zip”.)
                            The initial submission is v0. Resubmitted files
                            must have the version number incremented. 
 
                 
             
            - UGS NX3 
                
                    - The benchmark must be run using build “3.0.3.2” of
                        Unigraphics NX3. Submissions are only accepted on
                        Windows XP Professional SP2 and later. 
 
                    - The script file “%ProgramFiles%\SPECapc\nx3bench\nx3runbench.bat” may
                        be modified to change the directory where the benchmark
                        files are located on the system being tested. If
                        modified, the modified version must be included in
                        the benchmark submission. 
 
                    - The application window must be maximized and all
                        toolbars and the resource bar must be turned off. 
 
                    - NX3's Fixed Frame Rate feature must be disabled. 
 
                    - The system's booted configuration as specified in
                        the "boot.ini" file must contain the "/3GB" flag. 
 
                    - The chosen setting for View Frustum Culling must
                        be documented in the Comments field of the "results.txt" file. 
 
                    - The "Disable Line Smooth" option under
                        Preferences>Visualization Performance>General
                        Settings must be unchecked. 
 
                    - All other application settings should be default. 
 
                    - The submission must contain the “nx3result.txt” file
                        as well as the corresponding “nx3result.xls” file
                        generated from running the benchmark. Both of these
                        files may be found in the "results" directory
                        after a successful run of the benchmark. The first
                        section of the “nx3bench.xls” file must
                        be edited to reflect the system configuration. 
 
                    - The directory structure of the submission must be
                        as follows: 
 
                        .../company-name/system_1/nx3bench/nx3result.txt  
                        .../company-name/system_1/nx3bench/nx3bench.xls  
                        .../company-name/system_1/nx3bench/result.txt  
                        .../company-name/system_1/nx3bench/nx3bench.log  
                        .../company-name/system_1/nx3bench/nx3runbench.bat,if
                        modified, as required by rule c above)  
                        .../company-name/system_2/nx3bench/nx3result.txt  
                        .../company-name/system_2/nx3bench/nx3bench.xls  
                        .../company-name/system_2/nx3bench/result.txt  
                        .../company-name/system_2/nx3bench/nx3bench.log  
                        .../company-name/system_2/nx3bench/nx3runbench.bat,
                        if modified, as required by rule c above)  
                        etc...  
                    - The submission file must be named company_apc_nx3bench_vN.zip
                        or company_apc_nx3bench_vN.tar.z where company is
                        the member company or organization name in lower
                        case and vN is the file version (e.g. “sgi_apc_nx3bench_v0.tar.z” and “intel_apc_nx3bench_v0.zip”.)
                        The initial submission is v0. Resubmitted files must
                        have the version number incremented. 
 
                 
             
            - 3ds Max 8
                
                    -  The benchmark must be run using 3ds Max
                        version 8, service pack 3. 
 
                    - To ensure stability, the system must be booted with “/3GB” in
                        the relevant line of “boot.ini”. 
 
                    - The 3ds Max 8 window must be maximized, filling the
                        screen. The window must not be occluded by any other
                        windows. Note: Task-bar auto-hide must be enabled. 
 
                    - Drivers and custom drivers must be configured (explicitly
                        or implicitly) to use:
                        
                            -  GL_TEXTURE_MIN_FILTER = GL_LINEAR_MIPMAP_LINEAR 
 
                            - GL_TEXTURE_MAG_FILTER = GL_LINEAR 
 
                            - Textures cannot be modified, e.g. reduced
                                in size or depth. 
 
                            - Wireframe objects MUST be drawn using triangle
                                strips. (Note: this differs from the previous
                                version of the benchmark.) 
 
                         
                     
                    - The following Viewport parameters must be checked
                        (i.e. enabled) in the Viewports Tab found in the
                        3ds Max 8 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Backface cull on object creation 
 
                            - Mask viewport to safe region 
 
                            - Update background while playing 
 
                            - Display world axis 
 
                         
                     
                    - The following Viewport parameters must NOT be checked
                        in the 3ds max 8 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Attenuate lights 
 
                            - Filter environment backgrounds 
 
                         
                     
                    - The benchmark can be run using either OpenGL or Direct3D.
                        3ds Max 8 supports shaders only for Direct3D; therefore,
                        OpenGL submissions should contain a Shaders score
                        of "N/A". 
 
                    - For OpenGL
                            submissions, an application graphics driver or
                            plug-in may be used, but the images captured
                            from the benchmark must not differ from the supplied
                            3ds Max OpenGL driver's images by more than 5%.
                            If the application driver or plug-in’s
                            images do not match, the submitter may apply
                            for a waiver based on documented errors in the
                            3ds max OpenGL driver. The availability and support
                            of an application driver or plug-in shall not
                            affect the submission's “single supplier” or “parts
                            built” status. 
 
                    - The submission must contain the files “result.txt” and “result.xls”. 
 
                    - The directory structure of the submission must be
                        as follows:
 
                        .../company-name/system_1/3dsmax8/result.txt  
                        .../company-name/system_1/3dsmax8/result.xls  
                        .../company-name/system_2/3dsmax8/result.txt 
                        .../company-name/system_2/3dsmax8/result.xls  
                        etc...  
                    - The submission file must be named company_apc_3dsmax8_vN.zip
                        where company is the member company or organization
                        name in lower case and vN is the file version
                        (e.g. hp_apc_3dsmax8_v0.zip.) The initial submission
                        is v0. Resubmitted files must have the version number
                        incremented. 
 
                 
             
            - Solid Edge V19
                
                    -  The benchmark must be run using Solid
                            Edge V19.00.00.66 (SP0). 
 
                    - The application can be run
                            with Tools->Options->View Application Display
                            set to Graphics Card Driven (Basic or Advanced),
                            Software Driven or Backing Store. The choice
                            must be documented in the comments portion of
                            the results. 
 
                    - Application settings
                            must not be changed from the defaults set by
                            the benchmark installation. Settings that must
                            not be changed include, but are not limited to:
                        
                            - culling 
 
                            - wireframe
                                    display in Move Part command 
 
                            - arc smoothness
                                    (3) 
 
                         
                     
                    - The submission must
                            contain the file result.txt that is generated
                            during the benchmark run. 
 
                    - The directory structure
                            of the submission must be as follows:
 
                        .../company-name/system_1/SEV19/result.txt 
                        .../company-name/system_2/SEV19/result.txt  
                    - 
                        
The
                                submission file must be named company_apc_SolidEdgeV19_vN.zip
                                where company is the member company
                                or organization name in lower case and vN is
                                the file version (e.g. hp_apc_SolidEdgeV19_v0.zip.)
                                The initial submission is v0. Resubmitted
                                files must have the version number incremented.  
                     
                 
             
            - Solidworks 2007
                
                    - The benchmark must be run using Solidworks2007 service
                        pack 0. 
 
                    - The application window size must not be changed from
                        its initial size. 
 
                    - The submission must contain a result.txt file generated
                        by running the benchmark. Its contents are extracted
                        from the file apcresultsN.txt, generated by
                        the benchmark, where N is the number
                        of the benchmark test run. 
 
                    - The submission will be derived from the best composite
                        generated by the default 5 benchmark runs, as controlled
                        and reported by the benchmark GUI. 
 
                    - The appearance of the the application's Quick Tips
                        / Dynamic Help box, or other pop-ups that do not
                        dismiss themselves, will cause the benchmark run
                        to be invalid. The benchmark FAQ has information
                        on how to prevent this. 
 
                    - The directory structure of the submission must be
                        as follows:
 
                        .../company-name/system_1/sw2007/result.txt 
                        .../company-name/system_2/sw2007/result.txt 
                        etc...  
                    - The submission file must be named company_apc_sw2007_vN.zip
                        where company is the member company or
                        organization name in lower case and vN is
                        the file version (e.g. ibm_apc_sw2007_v0.zip.) The
                        initial submission is v0. Resubmitted files must
                        have the version number incremented. 
 
                 
             
            - 3ds Max 9 
                
                    -  The benchmark must be run using Autodesk
                        3ds Max version 9 with service pack SP2 applied. 
 
                    - To ensure stability, the system must be booted with “/3GB” in
                        the relevant line of “boot.ini”. 
 
                    - The 3ds Max 9 window must be maximized, filling the
                        screen. The window must not be occluded by any other
                        windows. Note: Task-bar auto-hide must be enabled. 
 
                    - Drivers and custom drivers must be configured (explicitly
                        or implicitly) to use:
                        
                            -  GL_TEXTURE_MIN_FILTER = GL_LINEAR_MIPMAP_LINEAR 
 
                            - GL_TEXTURE_MAG_FILTER = GL_LINEAR 
 
                            - Textures cannot be modified, e.g. reduced
                                in size or depth. 
 
                            - Wireframe objects must be drawn using triangle
                                strips. 
 
                         
                     
                    - The following Viewport parameters must be checked
                        (i.e. enabled) in the Viewports Tab found in the
                        3ds Max 9 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Backface cull on object creation 
 
                            - Mask viewport to safe region 
 
                            - Update background while playing 
 
                            - Display world axis 
 
                         
                     
                    - The following Viewport parameters must NOT be checked
                        in the 3ds max 9 Customize->Preferences menu (Viewports
                        tab):
                        
                            -  Attenuate lights 
 
                            - Filter environment backgrounds 
 
                         
                     
                    - The benchmark can be run using either OpenGL or DirectX.
                        3ds Max 9 supports shaders only for DirectX; therefore,
                        OpenGL submissions should contain a Shaders score
                        of "N/A". 
 
                    - For either
                            OpenGL or DirectX submissions an application
                            graphics driver or plug-in may be used, but the
                            images captured from the benchmark must not differ
                            from 3ds max's respective native OpenGL or DirectX
                            driver's images by more than 5%. If the application
                            driver's or plug-in’s images do not match,
                            the submitter may apply for a waiver based on
                            documented errors in the respective 3ds max native
                            driver. The availability and support of an application
                            driver or plug-in shall not affect the submission's “single
                            supplier” or “parts built” status. 
 
                    - The submission must contain the files “result.txt” and “result.xls”. 
 
                    - The directory structure of the submission must be
                        as follows:
 
                        .../company-name/system_1/3dsmax9/result.txt  
                        .../company-name/system_1/3dsmax9/result.xls  
                        .../company-name/system_2/3dsmax9/result.txt 
                        .../company-name/system_2/3dsmax9/result.xls  
                        etc...  
                    - The submission file must be named company_apc_3dsmax9_vN.zip
                        where company is the member company or organization
                        name in lower case and vN is the file version
                        (e.g. hp_apc_3dsmax9_v0.zip.) The initial submission
                        is v0. Resubmitted files must have the version number
                        incremented. 
 
                 
             
            - UGS NX4
                
                    - The benchmark must be run using build “Maintenance
                        Release 3” of Unigraphics NX4. Submissions
                        are accepted on Windows XP Professional SP2 and later,
                        and on Windows XP 64-bit Edition. 
 
                    - The script file “%ProgramFiles%\SPECapc\nx4mark\nx4mark.bat” may
                        be modified to change the directory where the benchmark
                        files are located on the system being tested. If
                        modified, the modified version must be included in
                        the benchmark submission. 
 
                    - The application window must be maximized and all
                        toolbars and the resource bar must be turned off. 
 
                    - NX4's Fixed Frame Rate feature must be disabled. 
 
                    - For 32-bit Windows, the system's booted configuration
                        as specified in the "boot.ini" file must
                        contain the "/3GB" flag. 
 
                    - The chosen setting for View Frustum Culling must
                        be documented in the Comments field of the "result.txt" file. 
 
                    - The "Disable Line Smooth" option under
                        Preferences>Visualization Performance>General
                        Settings must be unchecked. 
 
                    - All other application settings should be default. 
 
                    - The submission must contain the “nx4result.txt” file
                        as well as the corresponding “nx4result.xls” file
                        generated from running the benchmark. Both of these
                        files may be found in the "results" directory
                        after a successful run of the benchmark. The first
                        section of the “nx4bench.txt” file must
                        be edited to reflect the system configuration. 
 
                    - The directory structure of the submission must be
                        as follows: 
 
                        .../company-name/system_1/nx4mark/nx4mark_results.xls  
                        .../company-name/system_1/nx4mark/result.txt  
                        .../company-name/system_1/nx4mark/nx4mark.bat,
                        if modified, as required by rule c above)  
                         
                        .../company-name/system_1/nx4mark/nx4mark_results.xls  
                        .../company-name/system_1/nx4mark/result.txt  
                        .../company-name/system_1/nx4mark/nx4mark.bat,
                        if modified, as required by rule c above)  
                         
                        etc...  
                    - The submission file must be named company_apc_nx4mark_vN.zip
                        or company_apc_nx4mark_vN.tar.z where company is
                        the member company or organization name in lower
                        case and vN is the file version (e.g. “sun_apc_nx4mark_v0.tar.z” and “intel_apc_nx4mark_v0.zip”.)
                        The initial submission is v0. Resubmitted files must
                        have the version number incremented. 
 
                 
             
         
     
 
 
 
    - Adoption
 
         v2.10: Adopted 13 September 2007 (reflects GPC-to-GWPG transition,
        and adds NX4 rules) 
        v2.06: Adopted 13 July 2007 
        v2.05: Adopted 18 April 2007 
        v2.04: Adopted 06 February 2007 
        v2.03: Adopted  20 July 2006 
        v2.02: Adopted  27 April 2006 for 01 May 2006 publication 
        v2.01: Adopted  26 January 2006  
        v2.00: Adopted  26 January 2006  
 
  |