Solutions

Note: Applies to NDDS 3.x, RTI Connext 4.x and above. The LynxOS 4.0 IP stack (and many other BSD-derived network stacks) will fail to send multicast messages unless there is an entry in the routing table that matches the destination multicast IP address. In practice, this means that either: a ...
5373 reads — 0 comments
This solution addresses the following error message on Solaris systems: RTIOsapiSharedMemorySegment_attach:OS shmat() failure, error 0X18 NDDS_Transport_Shmem_attach_writer:failed to initialize shared memory resourcesegment for key 0x4020d0 Error message 0x18 on a Solaris system means that the ...
5564 reads — 0 comments
NOTE: Applies to RTI Data Distribution Service 4.1 and 4.2 By default, RTI Data Distribution Service uses shared memory to communicate with other DomainParticipants on the same node and disables local traffic over the UDPv4 or UDPv6 loopback interface. So, by default, an RTI Data Distribution ...
7259 reads — 0 comments
Note: Applies to RTI Connext 4.x and above RTI Connext 4.x and above uses a default multicast time-to-live (TTL) value of 1. This means that, by default, multicast traffic will not cross subnet boundaries. The TTL value can be adjusted in the transport property. Method One for modifying the TTL ( ...
16317 reads — 0 comments
Note: Applies to RTI Connext 4.x and above. An RTI Connext application installed as a Windows Service application (formerly known as an NT service) will not communicate over shared memory with a DDS application that is not installed as a Windows Service. This is due to the namespace used to create ...
7227 reads — 0 comments
Note: FlexLM is no longer a dependency for RTI Connext 4.5e and above. Fedora 7 lacks a library required by FlexLM. RTI Developer Platform uses FlexLM. You will encounter the following error during start up of RTI Developer Platform: License Error RTI Development Platform:failure to load ...
4364 reads — 0 comments
Note: Applies to RTI Data Distribution Service 4.4b and below If a network interface is disconnected or disabled while your application is running, your application might hang when properly shutting down your DomainParticipant. The likely cause is how the middleware cleans up sockets and how ...
4775 reads — 0 comments

Pages