Difference between revisions of "V4 Name Server"

From EPICSWIKI
m (Reduced to essentials)
 
(19 intermediate revisions by 2 users not shown)
Line 1: Line 1:
== Task ==
== Directory Server ==
Investigate a network client API for data and name resolution.


=== Name Server ===
EPICS ChannelAccess currently uses a broadcast mechanism for resolving
Hoping to use something existing. Get OpenLDAP [http://www.openldap.org] and check the following:
channel names.
* Can it be used to store record names, server addresses,  maybe more ('primary' server, fields, ...)? Try a simple setup, use the command-line ldapsearch to retrieve data.
Advantages:
* Does it allow for redundant servers, automatically syncing?
* Minimal if not zero initial setup.
* How to best interface it (API)?
* At least initially very fast, because a client can send multiple name requests in one network packet, all servers receive it simultaneously and can answer to known names.


=== Data ===
Disadvantages:
Is the low-level V4 CA client API all that's needed?
* Unclear how to extend this service to redundant servers, archived data, ...
Or is a wrapper needed that e.g. holds the data in a
* Currently, no wildcard searches.
container?
* Currently, no statistics on available number of channels, types of records, ...
* Waiting for 'caget' example code.
* Bigger installations run into problems with broadcast traffic:
** All devices see the requests, even though only one server has a given channel name.
** Continued searches for unresolved names to overcome UDP limitations.
 
As Larry Hoff pointed out, we really want a ''Directory'' server for
EPICS, something beyond a name server. A directory will not only map PV
names to IP addresses of ChannelAccess servers, but also support
* wildcard searches, reports on number of channels per IOC etc.
* redundant servers
* info about reduntant IOCs with the same PV, archived data for a PV
* backup/restore locations for a PV
* PV meta information like engineer-in-charge, CA security rules, ...
 
== Channels, Records, PVs, Properties ==
We currently have records with fields and channels with properties.
Each field of a record can become a channel name,
the field itself ends up in the 'value' property of the channel,
and the record decides how to fill the remaining properties.
 
Will the directory list all record names or all possible channels?
I think only record names.
 
Will it include the record type? Not enough information if we add user-defined fields.
Will it lists all fields of a record?
 
I think for now that only listing record names is enough in order to allow
connections.
 
The CA client would then have to query the CA server for all the available properties
of a PV, and note that properties might even change at runtime when we allow
online add & delete.
 
Whenever a tool like a command-line 'caget' or the 'probe' GUI requires a PV name,
it needs to be decided what properties "fred" would access. In the past, it was equal
to "fred.VAL". There is much more detail in the 'view' section of the
[[V4 DBD Statement Syntax]].
 
== LDAP, Directory, RDB, Schema, Entries, dn, cn, dc, ... ==
 
The ''Lightweight Directory Access Protocol'' (LDAP) is a network protocol.
It defines how to access data over the network, regardless of language
or underlying data storage. In contrast, RDBs like Oracle or MySQL define the
data storage and SQL dialect, while the network access to the data is then
added almost as an afterthought.
 
LDAP is meant as a Directory for simple attribute-based information
for fast, high volume lookup. Simple search filters are supported:
Patterns, "and", "or". In contrast to RDBs, there is no support for
arbitrary relations between entries, there are no transactions nor rollback.
 
There are open-source servers for LDAP, and client libraries for
C, Perl, Java and other languages. MacOSX, Linux, Windows 2000
already include LDAP clients. MS ActiveDirectory uses LDAP.
 
LDAP requires the data to be organized into hierarchical ''Entries''.
Each Entry has a globally unique ''Distinguished Name'' (dn),
and various ''Attributes'', that is Type/Value pairs.
The ''Schema'' defines which attributes an entry must have and which ones
it might have. There are existing schemas for DNS-type data or personell
information, including types like ''Common Name'' (cn) or ''Domain Component'' (dc),
but one can add custom schemata. Values might include binary data, e.g. images.
 
=== Mapping EPICS PV Information to LDAP ===
 
Each LDAP entry must have a unique name (dn). EPICS PVs must be unique as well,
but only within the search domain currently defined by the EPICS_CA_ADDR_LIST.
It might make sense to synchronize LDAP servers across these boundaries,
in which case the LDAP dn will have to include more than just the PV name.
For example, one could have a PV "fred" in both the accelerator and office network
of the SNS, and enter them into an SNS-wide LDAP directory like this:
 
fred.linac.sns.epics:  cas="linacioc1:5065"
fred.office.sns.epics: cas="testioc47.sns.ornl.gov:7987", description="Fred's test PV"
 
In this example, the directory includes the ChannelAccess server for both PVs,
and in one case also a description.
 
While an EPICS-specific tool might present the data as shown above,
most generic LDAP tools and APIs use the ''LDAP Data Interchange Format'' (LDIF).
In LDIF, part of the above data could look like this:
 
dn: dc=epics
objectclass: dcObject
objectclass: organization
o: Root of all EPICS PVs
dc: epics
 
dn: dc=sns,dc=epics
# Details omitted; similar to previous entry
 
dn: dc=linac,dc=sns,dc=epics
objectclass: dcObject
objectclass: organization
o: Root of all EPICS PVs in the SNS Linac
dc: linac
 
dn: cn=fred,dc=linac,dc=sns,dc=epics
objectclass: ProcessVariable
cn: fred
cas: linacioc1:5065
 
Some points I had to learn:
* Each entry has an objectclass. The schema for "ProcessVariable" would for example require a "cn" attribute, with optional "cas", "description" and maybe other attributes.<br>The schema would also determine if searches for PV names are case-sensitive or not.
* The hierarchical path "fred.linac.sns.epics" translates into a dn "cn=fred,dc=linac,dc=sns,dc=epics" which lists the relative dn of the entry itself and all path elements.
* Each path element must exists, so to bootstrap one has to create entries "dc=epics" and "dc=sns,dc=epics" etc., for which I used the pre-defined "organization" schema.
* To locate the PV 'fred', any of the following LDAP searches would work:
** Scope "base" with base "cn=fred,dc=linac,dc=sns,dc=epics".
** Scope "one" with base "dc=linac,dc=sns,dc=epics" and filter "(cn=fred)".
** Scope "one" with base "dc=linac,dc=sns,dc=epics" and filter "(cn=f*d)",<br>though this might return more records which happen to match the pattern.
** Scope "sub" with base "dc=sns,dc=epics" and filter "(cn=fred)",<br>which would also return an entry with dn "cn=fred,dc=office,dc=sns,dc=epics".
** There is a filter option "~=" for 'approximately equal'. Unclear what exactly defines approximate, but this might be very handy for detecing typos.
 
=== OpenLDAP ===
 
OpenLDAP  [http://www.openldap.org] is an open-source LDAP
client library for C and an LDAP server which can store the
actual data in /dev/null, a BerkeleyDB, SQL, Perl, a remote LDAP server.
 
OpenLDAP includes authentication and encrypted (SSL) transport,
replication (single master -> multiple slaves),
and referrals from one server to another one which has more details.
 
=== EPICS PV Tests ===
 
Using OpenLDAP-stable-20050429 and perl-ldap-0.3202 on a 1.3 GHz PowerBook G4
with 780 MB RAM under MacOS X 10.3.8, I used a perl script to create
PV entries, search them, and remove them.
In addition, the 'ldapsearch' that comes with RedHat Enterprise AS 3 was
used to check remote accessibillity.
 
When trying to list all PVs with an appropriate search pattern,
the default server configuration limits the response to 500 answers.
This was changed such that 'paged' searches which keep requesting data
in increments of e.g. 500 entries are allowed to continue until all data is retrieved.
From the command-line tool, that's done like this:
ldapsearch -x -LLL -E pr=10/noprompt -b 'dc=epics' '(cn=testpv*)'  cas
 
== Performance ==
 
Based on the default OpenLDAP server config with an additional
index "cn pres,eq" on the common name of the PVs,
PVs with names 'testrec0000000001', 'testrec0000000002' etc.
were added, then looked up one-by-one via their exact name,
and finially deleted one by one.
 
When using a C++ test client based on the OpenLDAP library,
the results for smaller numbers are:
Add        10000 records : 22.76 secs
Locate      10000 records : 6.49 secs
Delete      10000 records : 22.20 secs
 
In another test, a very rough estimated PV count of the SNS linac
was used:
Add        200000 records : 675.89 secs (38.74 CPU)
Locate    200000 records : 142.85 secs (27.91 CPU)
Delete    200000 records : 807.28 secs (37.28 CPU)
 
Those 200000 records use about 900MB of file space.
The results indicate that one can add 300 .. 440 records/sec,
locate 1400.. 1540 records/sec and delete 250..450 records/sec.
Without an index, the addition and deletion will be faster
at the expense of lookup times.
 
A typical SNS LLRF IOCs has 2300 records.
On startup, sending those to the directory server would take around 7 seconds.
A typical SNS LLRF overview screen has about 400 PVs,
and current connection times vary from less than a second to
several tens of seconds, depending on how well the search requests
reach the IOCs.
Resolving their CA servers would take about 0.3 seconds with LDAP,
which would then be followed by the time required to actually connect
to those servers.
 
In previous test using perl, the individual PV search actually used an LDIF
filter with exact match like 'testrec0000000001'. In addition, all PVs matching
a filter 'testrec*' were retrieved, and no index was used:
Adding  1000 records:  4 wallclock secs ( 2.29 usr +  0.14 sys =  2.43 CPU)
Locating 1000 records: 49 wallclock secs ( 4.10 usr +  0.21 sys =  4.31 CPU)
Match all        1000:  1 wallclock secs ( 1.18 usr +  0.02 sys =  1.20 CPU)
Delete  1000 records:  4 wallclock secs ( 1.35 usr +  0.14 sys =  1.49 CPU)
 
When locating individual records via the complete LDIF base without a filter,
the search is much faster:
Locating 1000 records:  5 wallclock secs ( 3.20 usr +  0.18 sys =  3.38 CPU)
 
When adding an "index cn pres,eq" to the server config, the results are the
same for both the filter and the base search case.
No difference observed when tripling the cachesize from the default of 1000.
The behavior scales linearly with the number of channels,
always a little over 200 additions or lookups per second:
Adding  10000 records: 47 wallclock secs (21.72 usr +  1.28 sys = 23.00 CPU)
Locating 10000 records: 44 wallclock secs (32.20 usr +  1.55 sys = 33.75 CPU)
Match all        10000: 15 wallclock secs (12.15 usr +  0.64 sys = 12.79 CPU)
Delete  10000 records: 38 wallclock secs (15.41 usr +  1.16 sys = 16.57 CPU)
 
Adding  50000 records: 243 wallclock secs (110.36 usr +  7.18 sys = 117.54 CPU)
Locating 50000 records: 226 wallclock secs (164.97 usr +  7.67 sys = 172.64 CPU)
Match all        50000: 80 wallclock secs (63.09 usr +  3.42 sys = 66.51 CPU)
Delete  50000 records: 219 wallclock secs (82.11 usr +  6.61 sys = 88.72 CPU)
 
Adding  200000 records: 1293 wallclock secs (483.65 usr + 33.20 sys = 516.85 CPU)
 
"index cn,dc pres,eq,approx,sub" takes longer to insert/delete
without improved search performance:
Adding  10000 records: 104 wallclock secs (26.25 usr +  1.67 sys = 27.92 CPU)
Locating 10000 records: 45 wallclock secs (32.86 usr +  1.55 sys = 34.41 CPU)
Match all        10000: 15 wallclock secs (12.31 usr +  0.47 sys = 12.78 CPU)
Delete  10000 records: 97 wallclock secs (17.89 usr +  1.34 sys = 19.23 CPU)
 
The CPU load is a rough 50/50 split between the perl and slapd processes.
The perl ldap library is 100% perl, including the network handling and
quite some data conversions from arrays into hashes and back,
so a pure C client might be a little faster.
 
Though an index slows down the insertion of PVs, it speeds up certain
retrieval methods, as probably desired for an EPICS directory server.
The search mechanism is based on round-trip requests.
Searching PVs one PV name at a time is slower than requesting all
PVs that match a pattern. Unfortunately, typical CA clients will
have to request data for specific PV names, not by pattern.
 
==== JOH: For comparison purposes, here is the performance of CA's built-in name resolution. ====
 
Testing with 100000 channels named joh:bill
channel connect test
Elapsed Per Item =  0.00001683 sec,    59427.3 Items per sec, 31.4 Mbps
Search tries per chan - mean=1.000930 std dev=0.030482 min=1.000000 max=2.000000
 
== API ==
[[V4 CA Interfaces]] already has a lookup interface for a name service
that uses asynchronous callbacks and a property catalog to return the results.
 
This is the interface I used for the test code. It is synchronous without configurable timeout (bad), but includes addition and removal of names (good), and the PVEntry lists the actual properties of an entry.
 
  /// Client-Interface to the EPICS directory
  class Directory
  {
  public:
    /// Constuctor
    Directory();
 
    /// Destructor (will disconnect)
    ~Directory();
 
    /// Connect to directory server
    /// Better: URL ala "LDAP://cn=Manager,dc=epics:secret@localhost:389"
    /// which would allow other services without changing the API?
    bool connect(const char *host = "localhost:389",
                const char *user = "cn=Manager,dc=epics",
                const char *password = "secret");
 
    /// Disconnect from directory server
    void disconnect();
 
    /// Directory entry for one PV
 
    /// Unclear how to best deal with this with regard to real-time OS
    /// where we don't want to allocate too much mem at runtime.
    /// This API allows caller to reuse the same 'PVEntry' over and over.
    static const size_t SLEN = 100;
    static const size_t MAXCAS  = 3;
    static const char PATH_SEP = '/';
    class PVEntry
    {
    public:
        char pv[SLEN];
        char path[SLEN];
        char description[SLEN];
        char cas[MAXCAS][SLEN];
    };
 
    bool Directory::add_pventry(const PVEntry *pv);
 
    /// Given pv->pv, fills remaining PVEntry information from directory
    bool get_pventry(PVEntry *pv);
 
    bool delete_pventry(const char *pv, const char *path);
};
 
=== To be investigated ===
 
* How to access it from vxWorks
* How does an LDAP server used for EPICS cooperate with existing LDAP servers for DNS, email etc. Should it be one and the same? Use special port numbers for 'EPICS' LDAP?
* API: What type of API would EPICS tools use? Whatever LDAP library they use? An 'EPICS wrapper' around LDAP? Gasper suggested to look at the Java JNDI API for inspiration.
* Common Database headaches:
** To simply get a  PV into LDAP, one has to check if this PV already exists and then either 'add' or 'modify'.
** What if a channel is no longer available? An IOC shutting down could remove PVs from the LDAP server, but an IOC that crashes won't.
* Replication: How to use it, how fast is it etc.
* Authentication and encryption: How hard is it to configure?

Latest revision as of 16:05, 7 July 2005

Directory Server

EPICS ChannelAccess currently uses a broadcast mechanism for resolving channel names. Advantages:

  • Minimal if not zero initial setup.
  • At least initially very fast, because a client can send multiple name requests in one network packet, all servers receive it simultaneously and can answer to known names.

Disadvantages:

  • Unclear how to extend this service to redundant servers, archived data, ...
  • Currently, no wildcard searches.
  • Currently, no statistics on available number of channels, types of records, ...
  • Bigger installations run into problems with broadcast traffic:
    • All devices see the requests, even though only one server has a given channel name.
    • Continued searches for unresolved names to overcome UDP limitations.

As Larry Hoff pointed out, we really want a Directory server for EPICS, something beyond a name server. A directory will not only map PV names to IP addresses of ChannelAccess servers, but also support

  • wildcard searches, reports on number of channels per IOC etc.
  • redundant servers
  • info about reduntant IOCs with the same PV, archived data for a PV
  • backup/restore locations for a PV
  • PV meta information like engineer-in-charge, CA security rules, ...

Channels, Records, PVs, Properties

We currently have records with fields and channels with properties. Each field of a record can become a channel name, the field itself ends up in the 'value' property of the channel, and the record decides how to fill the remaining properties.

Will the directory list all record names or all possible channels? I think only record names.

Will it include the record type? Not enough information if we add user-defined fields. Will it lists all fields of a record?

I think for now that only listing record names is enough in order to allow connections.

The CA client would then have to query the CA server for all the available properties of a PV, and note that properties might even change at runtime when we allow online add & delete.

Whenever a tool like a command-line 'caget' or the 'probe' GUI requires a PV name, it needs to be decided what properties "fred" would access. In the past, it was equal to "fred.VAL". There is much more detail in the 'view' section of the V4 DBD Statement Syntax.

LDAP, Directory, RDB, Schema, Entries, dn, cn, dc, ...

The Lightweight Directory Access Protocol (LDAP) is a network protocol. It defines how to access data over the network, regardless of language or underlying data storage. In contrast, RDBs like Oracle or MySQL define the data storage and SQL dialect, while the network access to the data is then added almost as an afterthought.

LDAP is meant as a Directory for simple attribute-based information for fast, high volume lookup. Simple search filters are supported: Patterns, "and", "or". In contrast to RDBs, there is no support for arbitrary relations between entries, there are no transactions nor rollback.

There are open-source servers for LDAP, and client libraries for C, Perl, Java and other languages. MacOSX, Linux, Windows 2000 already include LDAP clients. MS ActiveDirectory uses LDAP.

LDAP requires the data to be organized into hierarchical Entries. Each Entry has a globally unique Distinguished Name (dn), and various Attributes, that is Type/Value pairs. The Schema defines which attributes an entry must have and which ones it might have. There are existing schemas for DNS-type data or personell information, including types like Common Name (cn) or Domain Component (dc), but one can add custom schemata. Values might include binary data, e.g. images.

Mapping EPICS PV Information to LDAP

Each LDAP entry must have a unique name (dn). EPICS PVs must be unique as well, but only within the search domain currently defined by the EPICS_CA_ADDR_LIST. It might make sense to synchronize LDAP servers across these boundaries, in which case the LDAP dn will have to include more than just the PV name. For example, one could have a PV "fred" in both the accelerator and office network of the SNS, and enter them into an SNS-wide LDAP directory like this:

fred.linac.sns.epics:  cas="linacioc1:5065"
fred.office.sns.epics: cas="testioc47.sns.ornl.gov:7987", description="Fred's test PV"

In this example, the directory includes the ChannelAccess server for both PVs, and in one case also a description.

While an EPICS-specific tool might present the data as shown above, most generic LDAP tools and APIs use the LDAP Data Interchange Format (LDIF). In LDIF, part of the above data could look like this:

dn: dc=epics
objectclass: dcObject
objectclass: organization
o: Root of all EPICS PVs
dc: epics
dn: dc=sns,dc=epics
# Details omitted; similar to previous entry
dn: dc=linac,dc=sns,dc=epics
objectclass: dcObject
objectclass: organization
o: Root of all EPICS PVs in the SNS Linac
dc: linac
dn: cn=fred,dc=linac,dc=sns,dc=epics
objectclass: ProcessVariable
cn: fred
cas: linacioc1:5065 

Some points I had to learn:

  • Each entry has an objectclass. The schema for "ProcessVariable" would for example require a "cn" attribute, with optional "cas", "description" and maybe other attributes.
    The schema would also determine if searches for PV names are case-sensitive or not.
  • The hierarchical path "fred.linac.sns.epics" translates into a dn "cn=fred,dc=linac,dc=sns,dc=epics" which lists the relative dn of the entry itself and all path elements.
  • Each path element must exists, so to bootstrap one has to create entries "dc=epics" and "dc=sns,dc=epics" etc., for which I used the pre-defined "organization" schema.
  • To locate the PV 'fred', any of the following LDAP searches would work:
    • Scope "base" with base "cn=fred,dc=linac,dc=sns,dc=epics".
    • Scope "one" with base "dc=linac,dc=sns,dc=epics" and filter "(cn=fred)".
    • Scope "one" with base "dc=linac,dc=sns,dc=epics" and filter "(cn=f*d)",
      though this might return more records which happen to match the pattern.
    • Scope "sub" with base "dc=sns,dc=epics" and filter "(cn=fred)",
      which would also return an entry with dn "cn=fred,dc=office,dc=sns,dc=epics".
    • There is a filter option "~=" for 'approximately equal'. Unclear what exactly defines approximate, but this might be very handy for detecing typos.

OpenLDAP

OpenLDAP [1] is an open-source LDAP client library for C and an LDAP server which can store the actual data in /dev/null, a BerkeleyDB, SQL, Perl, a remote LDAP server.

OpenLDAP includes authentication and encrypted (SSL) transport, replication (single master -> multiple slaves), and referrals from one server to another one which has more details.

EPICS PV Tests

Using OpenLDAP-stable-20050429 and perl-ldap-0.3202 on a 1.3 GHz PowerBook G4 with 780 MB RAM under MacOS X 10.3.8, I used a perl script to create PV entries, search them, and remove them. In addition, the 'ldapsearch' that comes with RedHat Enterprise AS 3 was used to check remote accessibillity.

When trying to list all PVs with an appropriate search pattern, the default server configuration limits the response to 500 answers. This was changed such that 'paged' searches which keep requesting data in increments of e.g. 500 entries are allowed to continue until all data is retrieved. From the command-line tool, that's done like this:

ldapsearch -x -LLL -E pr=10/noprompt -b 'dc=epics' '(cn=testpv*)'  cas

Performance

Based on the default OpenLDAP server config with an additional index "cn pres,eq" on the common name of the PVs, PVs with names 'testrec0000000001', 'testrec0000000002' etc. were added, then looked up one-by-one via their exact name, and finially deleted one by one.

When using a C++ test client based on the OpenLDAP library, the results for smaller numbers are:

Add         10000 records : 22.76 secs
Locate      10000 records : 6.49 secs
Delete      10000 records : 22.20 secs

In another test, a very rough estimated PV count of the SNS linac was used:

Add        200000 records : 675.89 secs (38.74 CPU)
Locate     200000 records : 142.85 secs (27.91 CPU)
Delete     200000 records : 807.28 secs (37.28 CPU)

Those 200000 records use about 900MB of file space. The results indicate that one can add 300 .. 440 records/sec, locate 1400.. 1540 records/sec and delete 250..450 records/sec. Without an index, the addition and deletion will be faster at the expense of lookup times.

A typical SNS LLRF IOCs has 2300 records. On startup, sending those to the directory server would take around 7 seconds. A typical SNS LLRF overview screen has about 400 PVs, and current connection times vary from less than a second to several tens of seconds, depending on how well the search requests reach the IOCs. Resolving their CA servers would take about 0.3 seconds with LDAP, which would then be followed by the time required to actually connect to those servers.

In previous test using perl, the individual PV search actually used an LDIF filter with exact match like 'testrec0000000001'. In addition, all PVs matching a filter 'testrec*' were retrieved, and no index was used:

Adding   1000 records:  4 wallclock secs ( 2.29 usr +  0.14 sys =  2.43 CPU)
Locating 1000 records: 49 wallclock secs ( 4.10 usr +  0.21 sys =  4.31 CPU)
Match all        1000:  1 wallclock secs ( 1.18 usr +  0.02 sys =  1.20 CPU)
Delete   1000 records:  4 wallclock secs ( 1.35 usr +  0.14 sys =  1.49 CPU)

When locating individual records via the complete LDIF base without a filter, the search is much faster: Locating 1000 records: 5 wallclock secs ( 3.20 usr + 0.18 sys = 3.38 CPU)

When adding an "index cn pres,eq" to the server config, the results are the same for both the filter and the base search case. No difference observed when tripling the cachesize from the default of 1000. The behavior scales linearly with the number of channels, always a little over 200 additions or lookups per second:

Adding   10000 records: 47 wallclock secs (21.72 usr +  1.28 sys = 23.00 CPU)
Locating 10000 records: 44 wallclock secs (32.20 usr +  1.55 sys = 33.75 CPU)
Match all        10000: 15 wallclock secs (12.15 usr +  0.64 sys = 12.79 CPU)
Delete   10000 records: 38 wallclock secs (15.41 usr +  1.16 sys = 16.57 CPU)
Adding   50000 records: 243 wallclock secs (110.36 usr +  7.18 sys = 117.54 CPU)
Locating 50000 records: 226 wallclock secs (164.97 usr +  7.67 sys = 172.64 CPU)
Match all        50000: 80 wallclock secs (63.09 usr +  3.42 sys = 66.51 CPU)
Delete   50000 records: 219 wallclock secs (82.11 usr +  6.61 sys = 88.72 CPU)
Adding  200000 records: 1293 wallclock secs (483.65 usr + 33.20 sys = 516.85 CPU)

"index cn,dc pres,eq,approx,sub" takes longer to insert/delete without improved search performance:

Adding   10000 records: 104 wallclock secs (26.25 usr +  1.67 sys = 27.92 CPU)
Locating 10000 records: 45 wallclock secs (32.86 usr +  1.55 sys = 34.41 CPU)
Match all        10000: 15 wallclock secs (12.31 usr +  0.47 sys = 12.78 CPU)
Delete   10000 records: 97 wallclock secs (17.89 usr +  1.34 sys = 19.23 CPU)

The CPU load is a rough 50/50 split between the perl and slapd processes. The perl ldap library is 100% perl, including the network handling and quite some data conversions from arrays into hashes and back, so a pure C client might be a little faster.

Though an index slows down the insertion of PVs, it speeds up certain retrieval methods, as probably desired for an EPICS directory server. The search mechanism is based on round-trip requests. Searching PVs one PV name at a time is slower than requesting all PVs that match a pattern. Unfortunately, typical CA clients will have to request data for specific PV names, not by pattern.

JOH: For comparison purposes, here is the performance of CA's built-in name resolution.

Testing with 100000 channels named joh:bill
channel connect test
Elapsed Per Item =   0.00001683 sec,    59427.3 Items per sec, 31.4 Mbps
Search tries per chan - mean=1.000930 std dev=0.030482 min=1.000000 max=2.000000

API

V4 CA Interfaces already has a lookup interface for a name service that uses asynchronous callbacks and a property catalog to return the results.

This is the interface I used for the test code. It is synchronous without configurable timeout (bad), but includes addition and removal of names (good), and the PVEntry lists the actual properties of an entry.

 /// Client-Interface to the EPICS directory
 class Directory
 {
 public:
   /// Constuctor
   Directory();
   /// Destructor (will disconnect)
   ~Directory();
   /// Connect to directory server
   /// Better: URL ala "LDAP://cn=Manager,dc=epics:secret@localhost:389"
   /// which would allow other services without changing the API?
   bool connect(const char *host = "localhost:389",
                const char *user = "cn=Manager,dc=epics",
                const char *password = "secret");
   /// Disconnect from directory server
   void disconnect();
   /// Directory entry for one PV
   /// Unclear how to best deal with this with regard to real-time OS
   /// where we don't want to allocate too much mem at runtime.
   /// This API allows caller to reuse the same 'PVEntry' over and over.
   static const size_t SLEN = 100;
   static const size_t MAXCAS  = 3;
   static const char PATH_SEP = '/';
   class PVEntry
   {
   public:
       char pv[SLEN];
       char path[SLEN];
       char description[SLEN];
       char cas[MAXCAS][SLEN];
   };
   bool Directory::add_pventry(const PVEntry *pv);
   /// Given pv->pv, fills remaining PVEntry information from directory
   bool get_pventry(PVEntry *pv);
   bool delete_pventry(const char *pv, const char *path);

};

To be investigated

  • How to access it from vxWorks
  • How does an LDAP server used for EPICS cooperate with existing LDAP servers for DNS, email etc. Should it be one and the same? Use special port numbers for 'EPICS' LDAP?
  • API: What type of API would EPICS tools use? Whatever LDAP library they use? An 'EPICS wrapper' around LDAP? Gasper suggested to look at the Java JNDI API for inspiration.
  • Common Database headaches:
    • To simply get a PV into LDAP, one has to check if this PV already exists and then either 'add' or 'modify'.
    • What if a channel is no longer available? An IOC shutting down could remove PVs from the LDAP server, but an IOC that crashes won't.
  • Replication: How to use it, how fast is it etc.
  • Authentication and encryption: How hard is it to configure?