Go to primary content
Agile Product Lifecycle Management Capacity Planning Guide
Release 9.3.6
E71149-01
  Go To Table Of Contents
Contents

Previous
Previous
 
Next
Next
 

3 Breakpoint Testing Results

This chapter includes scalability data and breakpoint testing results.

3.1 Reports Scalability Data

This section provides results for reports scalability data.

3.1.1 Heap Consumption in Reports

When calculating resource needs, assume that reports will consume one-third of application server resources (Heap and CPU). For example, if the application server heap size is set to 1280, then the available heap on that application server for reports should be estimated as 426MB (1280m/3 = 426m).


Note:

Depending on the report, 1024 MB of available heap can typically handle up to 300K rows.

Total Heap Size (in megabytes) Available Heap Size for Reports Maximum rows based on available heap
1024 341 100,000
1280 427 125,000
2048 683 200,000
3072 1024 300,000

3.1.2 Protecting the Application Server

Agile PLM now provides two configuration parameters to limit report result set sizes and avoid out-of-memory errors. The configuration parameters are:

  • report.maxStandardBOMReportResults - set in the agile.properties file, this parameter sets the maximum number of BOM reports.

  • MaxReportResults - set in the Java Client, this parameter sets the maximum number of all non-BOM reports.

3.1.3 Example: Configuration Scenario

Scenario: The environment has a clustered deployment with 4 managed servers each having heap size of 3072m (assumes 64-bit JDK).

Considering the above scenario, the following table provides suggestions for the value of the report configuration parameters, depending on the desired number of reports to be executed concurrently.

Total Heap Size Maximum Result Set Per Node Desired Concurrent Reports Per Node Suggested Value For Above Parameters
3072m

3072m

300K rows 3 100K
3072m 300K rows 5 60K
3072m 300K rows 10 30K
3072m 300K rows 12 25K
3072m 300K rows 15 20K

3.2 Solution Results on Windows and Linux

The following tables show the results of breakpoint testing on the Product Collaboration, Product Quality Management, Product Portfolio Management, Product Cost Management, and Product Governance & Compliance solutions, and the Distributed File Manager and Export components. The testing was initially performed with Agile PLM running on Windows 2003 systems, but follow-up testing resulted in the same findings on Oracle Enterprise Linux 4 or RedHat Linux version 4


Note:

The response times for all of the tables are measured in seconds and are for single user only.

3.2.1 Product Collaboration

The following are breakpoint testing results for the Product Collaboration solution.

3.2.1.1 Change Object

The following table lists the results for Copy and Paste Affected Items.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
500 40 10% 10%
2000 240 25% 10%
3000 570 25% 10%
5000 Timed out 25% 12%

The following table lists the results for Save As Change.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 17 10% 5%
2500 57 15-25% 5%
5000 153 25% 5%
7500 268 25% 5%
10000 450 25% 5%

The following table lists the results for Change Status to Release (with Preliminary Items).

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 27 5% 10%
2500 70 5% 12%
5000 140 5% 12%
7500 214 5% 12%
10000 273 5% 12%

The following table lists the results for Change Status to Pending.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 42 5% 10%
2500 142 5% 10%
5000 305 25% 10%
7500 453 25% 10%

The following table lists the results for Change Status to Released. Each item has at least 3 pending changes.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 25 <5% 12%
2500 73 <5% 12%
5000 150 5% 10%
7500 220 <5% 12%

3.2.1.2 Item Object

The following table lists the results for Copy and Paste BOM.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
500 15 <5% <5%
2500 70 15-20% 5%
5000 150 15-20% 5%

The following table lists the results for Expand BOM-All Levels.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
3596 22 minutes 75% NA

Each page load took 5 seconds.

3.2.2 Product Quality Management

During the breakpoint testing for Product Quality Management, the table sizes used for testing purposes were as follows:

  • PSR - 23908 items

  • QCR - 23525 items

3.2.2.1 PSR Object

The following table lists the results for Copy and Paste Affected Items (Newly Imported Items).

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 52 10% 5%
2500 403 10-15% 5%
3500 720 10-15% 5%
5000 Timed out after 23 minutes 10-15% 5%

The following table lists the results for Copy and Paste Affected Items (Existing Items).

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
750 302 10% 5%
2500 Timed out 10-15% 5%

3.2.2.2 QCR Object

The following table lists the results for Copy and Paste Affected Items.

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database
2500 360 10-15% 5%
5000 Timed out after 22 min 10-15% 5%

3.2.3 Product Portfolio Management

The following table lists the results for several PPM scenarios.

Scenarios 1500 Object Tree(Response time in seconds) 3000 Object Tree(Response time in seconds)
Save As Project 30 50
Save As Template 18 35
Create Project From Template 38 95
Add 10 Team Members 21 42
Reschedule Root Project 13 30
Reschedule Child 11 24
Delegate Owner 15 27
Substitute Resource 11 24
Change Status 18 30
Expand All (Web Client) 160 NA

Tree Details:

  • 10 Team Members at each level

  • 10 Links

  • 10 Deliverables

  • 10 External Dependencies

3.2.4 Product Governance & Compliance

The following tables list the results for several PG&C scenarios.


Note:

Each declaration has 2 specs. Example: Dataset 1000=Declaration with 500 parts + 2 specs.

Table 3-1 Object: Declaration Tab Name: Items

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Items

12

25

35

15%

5%

View Items Tab

5

5

5

NA

NA

Calculate Compliance

4

4

4

NA

NA

Release the Declaration

40

72

150

NA

12%

Import Items (no BOS)

NA

NA

NA

NA

NA


Table 3-2 Object: Declaration Tab Name: Mfr Parts

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Mfr Parts

12

21

50

10-15%

5%

View Mfr Parts Tab

5

5

5

NA

NA

Calculate Compliance

4

4

4

NA

NA

Release the Declaration

40

75

90

NA

12%

Import Mfr Parts (no BOS)

NA

NA

NA

NA

NA


Table 3-3 Object: Declaration Tab Name: Part Groups

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Part Groups

12

22

50

10-15%

5%

View Part Groups Tab

5

5

5

NA

NA

Calculate Compliance

4

4

4

NA

NA

Release the Declaration

42

72

150

NA

12%

Import Part Groups (no BOS)

NA

NA

NA

NA

NA


Table 3-4 Object: Substance Groups Tab Name: Substances

Scenario 50 Substances 100 Substances

Search & Add Substances

6

8

View Substances tab

3

3

Import Substances

6

11


Table 3-5 Object: Specification Tab Name: Substances

Scenario 50 Substances 100 Substances

Search & Add Substances and/or Groups

6

8

View Substances tab

3

3

Import Substance Groups

5

9


3.2.4.1 Product Governance & Compliance Nightly Rollup

Table 3-6 Stats to Rollup 109462 BOMs with 1 Spec

Total Time Average App Server CPU Average Database CPU

14 Hours 30 Minutes

15-20%

12%



Note:

If there are more specifications, the rollups need to perform more processing. This also has an impact on performance.

3.2.5 Product Cost Management

The following table lists the results for several PCM scenarios.

Action Project with 2500 Items Project with 5000 Items BOM with 5000 Items Project with 10000 Items
Display project item tab 4 7 4 9
Search and Add Items into project from PC item master 41 65 22 NA
Update item from PC item master 31 48 70 117
Add/modify Partners (10) 25 45 28 73
Import item from external source 22 67 NA 160
Export item to external source 8 14 14 25
Project Save As 22 110 300 420
Quantity rollup 5 12 13 25
Display project AML tab 5 7 5 6
Display item BOM 5 7 4 9
Display of project analysis 5 6 6 8
Opening Responses Look Up 5 38 100 270
Project response price lookup 12 16 57 200
Set as best 21 45+55 (45 sec to open page + 55 to set as best) 80+95 110+138
Publish Quote History, Authoring 35+170 124+622 97+490 420 + Waited for 25 minutes
Publish Published Price Redlining 40+165 120+600 108+500 290+Out of process memory
Items RFQ Creation 30 153 110 306
Open RFQ 15 50 45 95
Display of RFQ Responses 15 21 10 24
Response Status Tab 3 8 6 7
Opening RFQ Response Look Up 12 38 105 260
RFQ Response Look Up 9 17 120 215
RFQ Response Export 18 56 57 107
RFQ Response Import 35 120 80 270
RFQ Response status display 3 8 6 7
Edit/Modify Responses 14 53 35 180
Supplier Response Export 19 45 35 87
Supplier Response Import 37 104 73 220
Submit Responses 19 53 35 104
Costed BOM Comparison (Buyer) Report 9 20 7 14
Delete Project 4 4 2 2
Close RFQ 14 45 2 90
Close RFQ (after Lock RFQ) 15 50 39 90
Delete RFQ 35 40 35 85
Lock RFQ 15 55 45 110
Analysis tab export 32 93 90 180
Data Sets:
  • Project with 2500 items, each having one AML

  • Project with 5000 items, each having two AMLs

  • Project with 10000 items, each having two AMLs

  • Five-level BOM with a total of 5000 items and 2 AMLs for each leaf component. The top-level assembly contains 10 immediate children with each child having 5 immediate children.


3.2.6 Distributed File Manager

The breakpoint testing for distributed file manager tested file upload and download on the distributed file management server. Basic file upload and download was tested on both supported application servers without a web proxy. Advanced file upload and download was tested using the Java Client on both supported application servers with a web proxy.

The following tables list the results for file upload and download.


Note:

Response time is measured in seconds.

Table 3-7 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 1 GB

96

NA

<5%

5-10%

Download - 1 GB

155

NA

25%

5-10%

Upload - 2 GB

Crashes

NA

NA

NA

Download - 2 GB

357

NA

10%

5%


The following table lists the results for advanced file upload and download.

Table 3-8 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 5GB

435

435

<5%

5%

Download -5GB

153

152

15%

10%


3.2.7 Export

Export was tested using the Web Client.

The following table lists the results for Export scenarios.

Table 3-9 Item/BOM Export Scenario

Scenario Data Set Response Time JVM Heap CPU Usage - App Server CPU Usage - DB Server Comments

Export CSV

4.8 Million Rows

4 h 30 mins

1200 MB

20% continuous

<5% continuous

6 level deep BOM

Output file size: 3 GB

Export PDX/aXML

4.8 Million Rows

25 mins

1200 MB

20% continuous

<5% continuous

Same BOM as above. Duplicated rows won't be exported to PDX and aXML formats.


Table 3-10 Changes Export Scenario

Scenario Data Set Response Time JVM Heap CPU Usage - App Server CPU Usage - DB Server Comments

Export CSV

3000 rows

3 h

1200 MB

20 %

<5%

NA


3.3 Solution Results on Solaris

The following tables show the results of breakpoint testing on the Product Collaboration, Product Quality Management, Product Portfolio Management, Product Cost Management solutions, and the Distributed File Manager component.

Breakpoint testing on the Solaris platform was performed on the following configuration:

  • Application Server: v490, IV+ Processor, 4 CPU - 8 Core, 16 GB RAM, 1.5 GHz, 32 MB L2 Cache

  • Database Server: v890, IV+ Processor, 8 CPU - 16 Core, 16 GB RAM, 1.5 GHz, 32 MB L2 Cache


Note:

The response times for all of the tables are measured in seconds and are for single user only.

3.3.1 Product Collaboration

This section provides breakpoint testing results for Product Collaboration.

3.3.1.1 Change Object

The following table lists the results for Copy and Paste Affected Items on Solaris.

Table 3-11 Results for Copy and Paste Affected Items on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

500

37

12%

<5%

2000

346

13%

<5%

3000

Timed out

14%

<5%


The following table lists the results for Save as Change on Solaris.

Table 3-12 Results for Save as Change on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

15

6%

1%

2500

45

15%

1%

5000

102

13%

1%

7500

210

15%

1%


The following table lists the results for Change Status to Released (with Preliminary Items) on Solaris.

Table 3-13 Results for Change Status to Released (with Preliminary Items) on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

7

7%

<5%

2500

20

12%

<5%

5000

44

10%

<5%

7500

62

13%

<5%


The following table lists the results for Change Status to Pending on Solaris.

Table 3-14 Results for Change Status to Pending on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

26

6%

<5%

2500

88

6%

<5%

5000

194

16%

<5%

7500

300

16%

<5%


The following table lists the results for Change Status to Released (each item has at least 3 pending changes) on Solaris.

Table 3-15 Results for Change Status to Released (each item has at least 3 pending changes) on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

9

4%

<5%

2500

22

6%

<5%

5000

46

11%

<5%


3.3.1.2 Item Object

The following table lists the results for Copy and Paste BOM on Solaris.

Table 3-16 Results for Copy and Paste BOM on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

500

11

12%

1%

2500

47

10%

1%

5000

110

12%

2%


The following table lists the results for Expand BOM (All Levels) on Solaris.

Table 3-17 Results for Expand BOM (All Levels) on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

3596

20

50%

NA


3.3.2 Product Quality Management

This section provides breakpoint testing results for Product Quality Management.

3.3.2.1 PSR Object

The following table lists the results for Copy and Paste Affected Items (newly imported items) on Solaris.

Table 3-18 Results for Copy and Paste Affected Items (newly imported items) on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

51

9%

1%

2500

451

10%

1%

3500

748

10%

1%

5000

Timed out

10%

1%


The following table lists the results for Copy and Paste Affected Items (existing items) on Solaris.

Table 3-19 Results for Copy and Paste Affected Items (existing items) on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

261

4%

1%

2500

Timed out

10%

1%


3.3.2.2 QCR Object

The following table lists the results for Copy and Paste Affected Items on Solaris.

Table 3-20 Results for Copy and Paste Affected Items on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

2500

390

9%

1%

5000

Timed Out

NA

NA


3.3.3 Product Portfolio Management

The following table lists the results of several PPM scenarios on Solaris.


Note:

Response times are in seconds.

Table 3-21 PPM Scenarios on Solaris

Scenarios 1200 Object Tree 2400 Object Tree

Save As Program

36

55

Save As Template

25

40

Create Project From Template

42

99

Add 10 Team Members

29

43

Reschedule Root Project

17

34

Reschedule Child

15

27

Delegate Owner

18

29

Substitute Resource

14

28

Change Status

22

39

Expand All

160

Out of memory


Tree Details:

  • 10 Team Members at each level

  • 10 Links

  • 10 Deliverables

  • 10 External Dependencies

The following tables list the results for several PG&C scenarios.


Note:

Each declaration has 2 specs. Example: Dataset 1000=Declaration with 500 parts + 2 specs.

Table 3-22 Object: Declaration Tab Name: Items

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Items

14

26

39

15%

5%

View Items Tab

4

4

4

NA

NA

Calculate Compliance

5

4

4

NA

NA

Release the Declaration

49

70

157

NA

12%

Import Items (no BOS)

NA

NA

NA

NA

NA


Table 3-23 Object: Declaration Tab Name: Mfr Parts

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Mfr Parts

12

22

52

10-15%

5%

View Mfr Parts Tab

4

6

6

NA

NA

Calculate Compliance

5

4

4

NA

NA

Release the Declaration

48

77

93

NA

12%

Import Mfr Parts (no BOS)

NA

NA

NA

NA

NA


Table 3-24 Object: Declaration Tab Name: Part Groups

Scenario 250 Compositions 500 Compositions 1000 Compositions App Server CPU Usage DB Server CPU Usage

Copy & Paste Part Groups

12

20

52

10-15%

5%

View Part Groups Tab

4

6

6

NA

NA

Calculate Compliance

4

4

4

NA

NA

Release the Declaration

49

75

156

NA

12%

Import Part Groups (no BOS)

NA

NA

NA

NA

NA


Table 3-25 Object: Substance Groups Tab Name: Substances

Scenario 50 Substances 100 Substances

Search & Add Substances

8

13

View Substances tab

3

3

Import Substances

8

14


Table 3-26 Object: Specification Tab Name: Substances

Scenario 50 Substances 100 Substances

Search & Add Substances and/or Groups

7

11

View Substances tab

3

3

Import Substance Groups

7

11


3.3.4 Product Cost Management

The following table lists the results for several PCM scenarios.


Note:

Response times are in seconds.

Action Project with 5000 Items Project with 10000 Items
Display project item tab 6 5
Search and Add Items into project from PC item master 57 NA
Update item from PC item master 45 194
Add/modify Partners (10) 32 60
Import item from external source 64 160
Export item to external source 13 29
Project Save As 210 545
Quantity rollup 17 25
Display project AML tab 4 6
Display item BOM 6 5
Display of project analysis 6 5
Opening Responses Look Up 14 30
Project response price lookup 14 26
Set as best 37+47 80+105
Publish Quote History, Authoring 47+45 163+75
Publish Published Price Redlining 46+52 110+70
Items RFQ Creation 45 110
Open RFQ 10 110
Display of RFQ Responses 4 6
Response Status Tab 4 6
Opening RFQ Response Look Up 10 16
RFQ Response Look Up 55 98
RFQ Response status display 4 6
Edit/Modify Responses 150 550
Supplier Response Export 135 480
Supplier Response Import 90 180
Submit Responses 88 510
Costed BOM Comparison (Buyer) Report 25 45
Delete Project 2 2
Close RFQ 270 690
Delete RFQ 30 33
Lock RFQ 291 690
Data Sets:
  • Project with 5000 items, each having two AMLs

  • Project with 10000 items, each having two AMLs


3.3.5 Distributed File Manager

The breakpoint testing for distributed file manager tested file upload and download on the distributed file management server. Basic file upload and download was tested on both supported application servers without a web proxy. Advanced file upload and download was tested using the Java Client on both supported application servers with a web proxy.

The following table lists the results for basic file upload and download.

Table 3-27 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 1 GB

93

NA

<5%

5%

Download - 1 GB

160

NA

25%

5%

Upload - 2 GB

Crashes

NA

NA

NA

Download - 2 GB

349

NA

10%

5%


The following table lists the results for advanced file upload and download.

Table 3-28 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 5GB

441

429

<5%

5%

Download -5GB

161

157

15%

5%


3.3.6 Export

Export was tested using the Web Client.

The following table lists the results for Export scenarios.

Table 3-29 Items/BOM Export Scenario

Scenario Data Set Response Time JVM Heap CPU Usage - App Server CPU Usage - DB Server Comments

Export CSV

4.8 Million Rows

4 h 30 mins

1200 MB

20% continuous

<5% continuous

6 level deep BOM

Output file size: 3 GB

Export PDX/aXML

4.8 Million Rows

25 mins

1200 MB

20% continuous

<5% continuous

Same BOM as above. Duplicated rows won't be exported to PDX and aXML formats.


Table 3-30 Changes Export Scenario

Scenario Data Set Response Time JVM Heap CPU Usage - App Server CPU Usage - DB Server Comments

Export CSV

3000 Rows

3 h

1200 MB

20%

<5%

NA


3.4 Solution Results on AIX

The following tables show the results of breakpoint testing on the Product Collaboration, Product Quality Management, and the Distributed File Manager component.

Breakpoint testing on the AIX platform was performed on the following configuration:

  • Application Server: PowerPC_Power6, P6 Processor, 4.7 GHz, 4 Cores, 16 GB RAM

  • Database Server: PowerPC_Power6, P6 Processor, 4.7 GHz, 4 Cores, 16 GB RAM


Note:

The response times for all of the tables are measured in seconds and are for single user only.

3.4.1 Product Collaboration

This section provides breakpoint testing results for Product Collaboration.

3.4.1.1 Change Object

The following table lists the results for Copy and Paste Affected Items on AIX.

Table 3-31 Results for Copy and Paste Affected Items on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

500

32

22%

5%

2000

255

25%

5%

3000

700

25%

5%

5000

Timed out after 15 min

25%

5%


The following table lists the results for Save as Change on AIX.

Table 3-32 Results for Save as Change on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

12

15%

5%

2500

41

15-25%

5%

5000

87

15-25%

5%

7500

168

15-25%

5%


The following table lists the results for Change Status to Released (with Preliminary Items) on AIX.

Table 3-33 Results for Change Status to Released (with Preliminary Items) on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

12

10%

5%

2500

20

10%

5%

5000

41

25%

5%

7500

70

25%

12%


The following table lists the results for Change Status to Pending on AIX.

Table 3-34 Results for Change Status to Pending on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

31

10%

6%

2500

85

25%

6%

5000

185

25%

6%

7500

300

25%

6%


The following table lists the results for Change Status to Released (each item has at least 3 pending changes) on AIX.

Table 3-35 Results for Change Status to Released on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

10

10%

10-15%

2500

21

10%

10-15%

5000

43

10%

10-15%


3.4.1.2 Item Object

The following table lists the results for Copy and Paste BOM on AIX.

Table 3-36 Results for Copy and Paste BOM on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

500

11

15%

<5%

2500

50

15-25%

<5%

5000

107

15-25%

5-10%


The following table lists the results for Expand BOM (All Levels) on AIX.

Table 3-37 Results for Expand BOM (All Levels) on AIX

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

3596

18

10%

<5%


3.4.2 Product Quality Management

This section provides breakpoint testing results for Product Quality Management.

3.4.2.1 PSR Object

The following table lists the results for Copy and Paste Affected Items (Newly Imported Items).

Table 3-38 Results or Copy and Paste Affected Items (Newly Imported Items)

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

48

10%

5%

2500

355

10%

5%

3500

710

10%

5%

5000

Timed out after 15 minutes

10%

5%


The following table lists the results for Copy and Paste Affected Items (Existing Items).

Table 3-39 Results for Copy and Paste Affected Items (Existing Items)

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

750

292

5%

2%

2500

Timed out after 15 minutes

NA

NA


3.4.2.2 QCR Object

The following table lists the results for Copy and Paste Affected Items on Solaris.

Table 3-40 Results for Copy and Paste Affected Items on Solaris

Number of Affected Items Response Time (in seconds) CPU Usage - Application Server CPU Usage - Database

2500

320

10%

5%

5000

Timed Out after 15 minutes

NA

NA


3.4.3 Distributed File Manager

The breakpoint testing for distributed file manager tested file upload and download on the distributed file management server. Basic file upload and download was tested on both supported application servers without a web proxy. Advanced file upload and download was tested using the Java Client on both supported application servers with a web proxy.

The following table lists the results for basic file upload and download.

Table 3-41 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 1 GB

45

NA

11%

10%

Download - 1 GB

10

NA

15%

5%

Upload - 2 GB

86

NA

11%

5%

Download - 2 GB

30

NA

50%

5%


The following table lists the results for advanced file upload and download.

Table 3-42 Basic File Upload and Download

Scenario Response Time - Web Client Response Time - Java Client CPU Usage - Client CPU Usage - DFM Server

Upload - 7 GB

205

255

10%

12%

Download - 7 GB

160

156

15-25%

20%


3.4.4 Export

Export was tested using the Web Client.

The following table lists the results for Export scenarios.

Table 3-43 Changes Export Scenario

Scenario Data Set Response Time CPU Usage - App Server CPU Usage - DB Server

Export - CSV

3500

30

<10%

<2%

Export - PDX

3500

35

<10%

<2%