Hello,
I am trying to run the following test:
start cmd /k perftest_java.bat -pub -dataLen 100 -domain 2 -latencyCount 1000 -numIter 1000000 -bestEffort -sendQueueSize 1 -pidMultiPubTest 0 -numSubscribers 1
start cmd /k perftest_java.bat -pub -dataLen 100 -domain 2 -latencyCount 1000 -numIter 1000000 -bestEffort -sendQueueSize 1 -pidMultiPubTest 1 -numSubscribers 1
start cmd /k perftest_java.bat -pub -dataLen 100 -domain 2 -latencyCount 1000 -numIter 1000000 -bestEffort -sendQueueSize 1 -pidMultiPubTest 2 -numSubscribers 1
start cmd /k perftest_java.bat -pub -dataLen 100 -domain 2 -latencyCount 1000 -numIter 1000000 -bestEffort -sendQueueSize 1 -pidMultiPubTest 3 -numSubscribers 1
start cmd /k perftest_java.bat -pub -dataLen 100 -domain 2 -latencyCount 1000 -numIter 1000000 -bestEffort -sendQueueSize 1 -pidMultiPubTest 4 -numSubscribers 1
start cmd /k perftest_java.bat -sub -dataLen 100 -domain 2 -bestEffort -numPublishers 5
reader_listener.print_summary_latency();
reader_listener = new LatencyListener(num_latency,_latencyTest?writer:null);
Attachment | Size |
---|---|
errar.png | 15.93 KB |
Hi Blitz3r,
Thanks for reporting this, and also for debugging the root cause. When using RTI Perftest to do a N-1 test (many to 1), latency is always reported by the Perftest Publisher with "pid" 0. This does not mean that the scenario you are executing is not supported, it just means that we will just answer to the 1st Publisher Application, by design.
Now, wrt your error, what version of RTI Perftest are you using? I mention this because that statement is, in the latest versions of Perftest, surrounded by this:
I also took the chance to review the test you are performing: I would like to ask:
Why are you setting the sendQueueSize to that number? That is going to slow down your system most likely. What kind of tests do you want to do? What are you measuring and what is the purpose of the test? I see you have been asking a bunch of questions about performance, perftest and particularly about the Java api for RTI Perftest. Could you let us know the use case (cases) you are interested in so O can somehow guide you?
Regards,
Javi
Hi Javier,
Apologies for the lateness of this response. I had surgery recently so I was and still am recovering. Thank you for taking the time to answer my post.
Here are my answers to your questions (please excuse the misnumbering):
Here are some examples of tests that I have run:
Test Set 1 [Throughput Test + Varied Latency Count]:
Constants: 500,000 Iterations + 1 Publisher + 10,000 Latency Count
Variables: 1/5/10 Subscribers
Test Set 2 [Best Effort VS Reliable]:
Constants: 1 Publisher + 1/5/10 Subscribers + 10,000 Latency Count + 500,000 Iterations
Variables: Best Effort/Reliable + Unicast/Multicast
Test Set 3 [100 VS 512 Bytes]:
Constants: 1 Publisher + 1/5/10 Subscribers + Best Effort Unicast + Best Effort Multicast + Reliable Unicast + Reliable Multicast
Variables: 100/512 Bytes
Test Set 4 [Instances]:
Constants: 1 Publisher + 1/5/10 Subscribers + 10 Instances
Variables: Best Effort/Reliable + Unicast/Multicast
I would like to run the following in the future:
Test Set 1 [Multipub 100 Byte]
Constants: 100 Bytes
Variables: 1/5/10 Publishers + 1/5/10 Subscribers + Best Effort/Reliable + Unicast/Multicast
Test Set 2 [Multipub 512 Byte]
Constants: 512 Bytes
Variables: 1/5/10 Publishers + 1/5/10 Subscribers + Best Effort/Reliable + Unicast/Multicast
Test Set 3 [Multipub 100 Byte]
Constants: 100 Bytes
Variables: 20/50/100 Publishers + 20/50/100 Subscribers + Best Effort/Reliable + Unicast/Multicast
Your guidance would be much appreciated and I look forward to hearing from you soon!
Thanks,
Blitz3r