summaryrefslogtreecommitdiff
path: root/libexec/rpc.lockd/procs.c
diff options
context:
space:
mode:
Diffstat (limited to 'libexec/rpc.lockd/procs.c')
-rw-r--r--libexec/rpc.lockd/procs.c588
1 files changed, 588 insertions, 0 deletions
diff --git a/libexec/rpc.lockd/procs.c b/libexec/rpc.lockd/procs.c
new file mode 100644
index 00000000000..54e5f814ceb
--- /dev/null
+++ b/libexec/rpc.lockd/procs.c
@@ -0,0 +1,588 @@
+/*
+ * Copyright (c) 1995
+ * A.R. Gordon (andrew.gordon@net-tel.co.uk). All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. All advertising materials mentioning features or use of this software
+ * must display the following acknowledgement:
+ * This product includes software developed for the FreeBSD project
+ * 4. Neither the name of the author nor the names of any co-contributors
+ * may be used to endorse or promote products derived from this software
+ * without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY ANDREW GORDON AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ *
+ */
+
+
+
+#include "lockd.h"
+
+#include <sys/param.h> /* for MAXHOSTNAMELEN */
+#include <sys/socket.h>
+#include <netinet/in.h>
+#include <arpa/inet.h>
+#include <netdb.h>
+
+
+#define CLIENT_CACHE_SIZE 64 /* No. of client sockets cached */
+#define CLIENT_CACHE_LIFETIME 120 /* In seconds */
+
+
+/* log_from_addr ----------------------------------------------------------- */
+/*
+ Purpose: Log name of function called and source address
+ Returns: Nothing
+ Notes: Extracts the source address from the transport handle
+ passed in as part of the called procedure specification
+*/
+
+static void log_from_addr(char *fun_name, struct svc_req *req)
+{
+ struct sockaddr_in *addr;
+ struct hostent *host;
+ char hostname_buf[40];
+
+ addr = svc_getcaller(req->rq_xprt);
+ host = gethostbyaddr((char *)&(addr->sin_addr), addr->sin_len, AF_INET);
+ if (host)
+ {
+ strncpy(hostname_buf, host->h_name, sizeof(hostname_buf));
+ hostname_buf[sizeof(hostname_buf) -1] = '\0';
+ }
+ else /* No hostname available - print raw address */
+ {
+ strcpy(hostname_buf, inet_ntoa(addr->sin_addr));
+ }
+
+ syslog(LOG_DEBUG, "%s from %s", fun_name, hostname_buf);
+}
+
+
+/* get_client -------------------------------------------------------------- */
+/*
+ Purpose: Get a CLIENT* for making RPC calls to lockd on given host
+ Returns: CLIENT* pointer, from clnt_udp_create, or NULL if error
+ Notes: Creating a CLIENT* is quite expensive, involving a
+ conversation with the remote portmapper to get the
+ port number. Since a given client is quite likely
+ to make several locking requests in succession, it is
+ desirable to cache the created CLIENT*.
+
+ Since we are using UDP rather than TCP, there is no cost
+ to the remote system in keeping these cached indefinitely.
+ Unfortunately there is a snag: if the remote system
+ reboots, the cached portmapper results will be invalid,
+ and we will never detect this since all of the xxx_msg()
+ calls return no result - we just fire off a udp packet
+ and hope for the best.
+
+ We solve this by discarding cached values after two
+ minutes, regardless of whether they have been used
+ in the meanwhile (since a bad one might have been used
+ plenty of times, as the host keeps retrying the request
+ and we keep sending the reply back to the wrong port).
+
+ Given that the entries will always expire in the order
+ that they were created, there is no point in a LRU
+ algorithm for when the cache gets full - entries are
+ always re-used in sequence.
+*/
+
+static CLIENT *clnt_cache_ptr[CLIENT_CACHE_SIZE];
+static long clnt_cache_time[CLIENT_CACHE_SIZE]; /* time entry created */
+static struct in_addr clnt_cache_addr[CLIENT_CACHE_SIZE];
+static int clnt_cache_next_to_use = 0;
+
+static CLIENT *get_client(struct sockaddr_in *host_addr)
+{
+ CLIENT *client;
+ int sock_no;
+ struct timeval retry_time, time_now;
+ int i;
+
+ gettimeofday(&time_now, NULL);
+
+ /* Search for the given client in the cache, zapping any expired */
+ /* entries that we happen to notice in passing. */
+ for (i = 0; i < CLIENT_CACHE_SIZE; i++)
+ {
+ client = clnt_cache_ptr[i];
+ if (client &&
+ ((clnt_cache_time[i] + CLIENT_CACHE_LIFETIME) < time_now.tv_sec))
+ {
+ /* Cache entry has expired. */
+ if (debug_level > 3) syslog(LOG_DEBUG, "Expired CLIENT* in cache");
+ clnt_cache_time[i] = 0L;
+ clnt_destroy(client);
+ clnt_cache_ptr[i] = NULL;
+ client = NULL;
+ }
+
+ if (client && !memcmp(&clnt_cache_addr[i], &host_addr->sin_addr,
+ sizeof(struct in_addr)))
+ {
+ /* Found it! */
+ if (debug_level > 3) syslog(LOG_DEBUG, "Found CLIENT* in cache");
+ return (client);
+ }
+ }
+
+ /* Not found in cache. Free the next entry if it is in use */
+ if (clnt_cache_ptr[clnt_cache_next_to_use])
+ {
+ clnt_destroy(clnt_cache_ptr[clnt_cache_next_to_use]);
+ clnt_cache_ptr[clnt_cache_next_to_use] = NULL;
+ }
+
+ /* Create the new client handle */
+
+ sock_no = RPC_ANYSOCK;
+ retry_time.tv_sec = 5;
+ retry_time.tv_usec = 0;
+ host_addr->sin_port = 0; /* Force consultation with portmapper */
+ client = clntudp_create(host_addr, NLM_PROG, NLM_VERS, retry_time, &sock_no);
+ if (!client)
+ {
+ syslog(LOG_ERR, clnt_spcreateerror("clntudp_create"));
+ syslog(LOG_ERR, "Unable to return result to %s",
+ inet_ntoa(host_addr->sin_addr));
+ return NULL;
+ }
+
+ /* Success - update the cache entry */
+ clnt_cache_ptr[clnt_cache_next_to_use] = client;
+ clnt_cache_addr[clnt_cache_next_to_use] = host_addr->sin_addr;
+ clnt_cache_time[clnt_cache_next_to_use] = time_now.tv_sec;
+ if (++clnt_cache_next_to_use > CLIENT_CACHE_SIZE)
+ clnt_cache_next_to_use = 0;
+
+ /* Disable the default timeout, so we can specify our own in calls */
+ /* to clnt_call(). [note that the timeout is a different concept */
+ /* from the retry period set in clnt_udp_create() above.] */
+ retry_time.tv_sec = -1;
+ retry_time.tv_usec = -1;
+ clnt_control(client, CLSET_TIMEOUT, &retry_time);
+
+ if (debug_level > 3) syslog(LOG_DEBUG, "Created CLIENT* for %s",
+ inet_ntoa(host_addr->sin_addr));
+ return client;
+}
+
+
+/* transmit_result --------------------------------------------------------- */
+/*
+ Purpose: Transmit result for nlm_xxx_msg pseudo-RPCs
+ Returns: Nothing - we have no idea if the datagram got there
+ Notes: clnt_call() will always fail (with timeout) as we are
+ calling it with timeout 0 as a hack to just issue a datagram
+ without expecting a result
+*/
+
+static void transmit_result(int opcode, nlm_res *result, struct svc_req *req)
+{
+ static char dummy;
+ struct sockaddr_in *addr;
+ CLIENT *cli;
+ int success;
+ struct timeval timeo;
+
+ addr = svc_getcaller(req->rq_xprt);
+ if (cli = get_client(addr))
+ {
+ timeo.tv_sec = 0; /* No timeout - not expecting response */
+ timeo.tv_usec = 0;
+
+ success = clnt_call(cli, opcode, xdr_nlm_res, result, xdr_void,
+ &dummy, timeo);
+
+ if (debug_level > 2) syslog(LOG_DEBUG, "clnt_call returns %d\n", success);
+ }
+}
+
+/* ------------------------------------------------------------------------- */
+/*
+ Functions for Unix<->Unix locking (ie. monitored locking, with rpc.statd
+ involved to ensure reclaim of locks after a crash of the "stateless"
+ server.
+
+ These all come in two flavours - nlm_xxx() and nlm_xxx_msg().
+ The first are standard RPCs with argument and result.
+ The nlm_xxx_msg() calls implement exactly the same functions, but
+ use two pseudo-RPCs (one in each direction). These calls are NOT
+ standard use of the RPC protocol in that they do not return a result
+ at all (NB. this is quite different from returning a void result).
+ The effect of this is to make the nlm_xxx_msg() calls simple unacknowledged
+ datagrams, requiring higher-level code to perform retries.
+
+ Despite the disadvantages of the nlm_xxx_msg() approach (some of which
+ are documented in the comments to get_client() above), this is the
+ interface used by all current commercial NFS implementations
+ [Solaris, SCO, AIX etc.]. This is presumed to be because these allow
+ implementations to continue using the standard RPC libraries, while
+ avoiding the block-until-result nature of the library interface.
+
+ No client implementations have been identified so far that make use
+ of the true RPC version (early SunOS releases would be a likely candidate
+ for testing).
+*/
+
+
+/* nlm_test ---------------------------------------------------------------- */
+/*
+ Purpose: Test whether a specified lock would be granted if requested
+ Returns: nlm_granted (or error code)
+ Notes:
+*/
+
+nlm_testres *nlm_test_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
+{
+ static nlm_testres res;
+ if (debug_level) log_from_addr("nlm_test", rqstp);
+
+ /* Copy the cookie from the argument into the result. Note that this */
+ /* is slightly hazardous, as the structure contains a pointer to a */
+ /* malloc()ed buffer that will get freed by the caller. However, the */
+ /* main function transmits the result before freeing the argument */
+ /* so it is in fact safe. */
+ res.cookie = arg->cookie;
+ res.stat.stat = nlm_granted;
+ return (&res);
+}
+
+void *nlm_test_msg_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
+{
+ nlm_testres res;
+ static char dummy;
+ struct sockaddr_in *addr;
+ CLIENT *cli;
+ int success;
+ struct timeval timeo;
+
+ if (debug_level) log_from_addr("nlm_test_msg", rqstp);
+
+ res.cookie = arg->cookie;
+ res.stat.stat = nlm_granted;
+
+ /* nlm_test has different result type to the other operations, so */
+ /* can't use transmit_result() in this case */
+ addr = svc_getcaller(rqstp->rq_xprt);
+ if (cli = get_client(addr))
+ {
+ timeo.tv_sec = 0; /* No timeout - not expecting response */
+ timeo.tv_usec = 0;
+
+ success = clnt_call(cli, NLM_TEST_RES, xdr_nlm_testres, &res, xdr_void,
+ &dummy, timeo);
+
+ if (debug_level > 2) syslog(LOG_DEBUG, "clnt_call returns %d\n", success);
+ }
+ return (NULL);
+}
+
+/* nlm_lock ---------------------------------------------------------------- */
+/*
+ Purposes: Establish a lock
+ Returns: granted, denied or blocked
+ Notes: *** grace period support missing
+*/
+
+nlm_res *nlm_lock_1_svc(nlm_lockargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_lock", rqstp);
+
+ /* copy cookie from arg to result. See comment in nlm_test_1() */
+ res.cookie = arg->cookie;
+
+ res.stat.stat = nlm_granted;
+ return (&res);
+}
+
+void *nlm_lock_msg_1_svc(nlm_lockargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+
+ if (debug_level) log_from_addr("nlm_lock_msg", rqstp);
+
+ res.cookie = arg->cookie;
+ res.stat.stat = nlm_granted;
+ transmit_result(NLM_LOCK_RES, &res, rqstp);
+
+ return (NULL);
+}
+
+/* nlm_cancel -------------------------------------------------------------- */
+/*
+ Purpose: Cancel a blocked lock request
+ Returns: granted or denied
+ Notes:
+*/
+
+nlm_res *nlm_cancel_1_svc(nlm_cancargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_cancel", rqstp);
+
+ /* copy cookie from arg to result. See comment in nlm_test_1() */
+ res.cookie = arg->cookie;
+
+ /* Since at present we never return 'nlm_blocked', there can never be */
+ /* a lock to cancel, so this call always fails. */
+ res.stat.stat = nlm_denied;
+ return (&res);
+}
+
+void *nlm_cancel_msg_1_svc(nlm_cancargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_cancel_msg", rqstp);
+
+ res.cookie = arg->cookie;
+ /* Since at present we never return 'nlm_blocked', there can never be */
+ /* a lock to cancel, so this call always fails. */
+ res.stat.stat = nlm_denied;
+ transmit_result(NLM_CANCEL_RES, &res, rqstp);
+ return (NULL);
+}
+
+/* nlm_unlock -------------------------------------------------------------- */
+/*
+ Purpose: Release an existing lock
+ Returns: Always granted, unless during grace period
+ Notes: "no such lock" error condition is ignored, as the
+ protocol uses unreliable UDP datagrams, and may well
+ re-try an unlock that has already succeeded.
+*/
+
+nlm_res *nlm_unlock_1_svc(nlm_unlockargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_unlock", rqstp);
+
+ res.stat.stat= nlm_granted;
+ res.cookie = arg->cookie;
+
+ return (&res);
+}
+
+void *nlm_unlock_msg_1_svc(nlm_unlockargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_unlock_msg", rqstp);
+
+ res.stat.stat = nlm_granted;
+ res.cookie = arg->cookie;
+
+ transmit_result(NLM_UNLOCK_RES, &res, rqstp);
+ return (NULL);
+}
+
+/* ------------------------------------------------------------------------- */
+/*
+ Client-side pseudo-RPCs for results. Note that for the client there
+ are only nlm_xxx_msg() versions of each call, since the 'real RPC'
+ version returns the results in the RPC result, and so the client
+ does not normally receive incoming RPCs.
+
+ The exception to this is nlm_granted(), which is genuinely an RPC
+ call from the server to the client - a 'call-back' in normal procedure
+ call terms.
+*/
+
+/* nlm_granted ------------------------------------------------------------- */
+/*
+ Purpose: Receive notification that formerly blocked lock now granted
+ Returns: always success ('granted')
+ Notes:
+*/
+
+nlm_res *nlm_granted_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_granted", rqstp);
+
+ /* copy cookie from arg to result. See comment in nlm_test_1() */
+ res.cookie = arg->cookie;
+
+ res.stat.stat = nlm_granted;
+ return (&res);
+}
+
+void *nlm_granted_msg_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
+{
+ nlm_res res;
+ if (debug_level) log_from_addr("nlm_granted_msg", rqstp);
+
+ res.cookie = arg->cookie;
+ res.stat.stat = nlm_granted;
+ transmit_result(NLM_GRANTED_RES, &res, rqstp);
+ return (NULL);
+}
+
+/* nlm_test_res ------------------------------------------------------------ */
+/*
+ Purpose: Accept result from earlier nlm_test_msg() call
+ Returns: Nothing
+*/
+
+void *nlm_test_res_1_svc(nlm_testres *arg, struct svc_req *rqstp)
+{
+ if (debug_level) log_from_addr("nlm_test_res", rqstp);
+ return (NULL);
+}
+
+/* nlm_lock_res ------------------------------------------------------------ */
+/*
+ Purpose: Accept result from earlier nlm_lock_msg() call
+ Returns: Nothing
+*/
+
+void *nlm_lock_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
+{
+ if (debug_level) log_from_addr("nlm_lock_res", rqstp);
+
+ return (NULL);
+}
+
+/* nlm_cancel_res ---------------------------------------------------------- */
+/*
+ Purpose: Accept result from earlier nlm_cancel_msg() call
+ Returns: Nothing
+*/
+
+void *nlm_cancel_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
+{
+ if (debug_level) log_from_addr("nlm_cancel_res", rqstp);
+ return (NULL);
+}
+
+/* nlm_unlock_res ---------------------------------------------------------- */
+/*
+ Purpose: Accept result from earlier nlm_unlock_msg() call
+ Returns: Nothing
+*/
+
+void *nlm_unlock_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
+{
+ if (debug_level) log_from_addr("nlm_unlock_res", rqstp);
+ return (NULL);
+}
+
+/* nlm_granted_res --------------------------------------------------------- */
+/*
+ Purpose: Accept result from earlier nlm_granted_msg() call
+ Returns: Nothing
+*/
+
+void *nlm_granted_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
+{
+ if (debug_level) log_from_addr("nlm_granted_res", rqstp);
+ return (NULL);
+}
+
+/* ------------------------------------------------------------------------- */
+/*
+ Calls for PCNFS locking (aka non-monitored locking, no involvement
+ of rpc.statd).
+
+ These are all genuine RPCs - no nlm_xxx_msg() nonsense here.
+*/
+
+
+/* nlm_share --------------------------------------------------------------- */
+/*
+ Purpose: Establish a DOS-style lock
+ Returns: success or failure
+ Notes: Blocking locks are not supported - client is expected
+ to retry if required.
+*/
+
+nlm_shareres *nlm_share_3_svc(nlm_shareargs *arg, struct svc_req *rqstp)
+{
+ static nlm_shareres res;
+ if (debug_level) log_from_addr("nlm_share", rqstp);
+
+ res.cookie = arg->cookie;
+ res.stat = nlm_granted;
+ res.sequence = 1234356; /* X/Open says this field is ignored? */
+ return (&res);
+}
+
+/* nlm_unshare ------------------------------------------------------------ */
+/*
+ Purpose: Release a DOS-style lock
+ Returns: nlm_granted, unless in grace period
+ Notes:
+*/
+
+nlm_shareres *nlm_unshare_3_svc(nlm_shareargs *arg, struct svc_req *rqstp)
+{
+ static nlm_shareres res;
+ if (debug_level) log_from_addr("nlm_unshare", rqstp);
+
+ res.cookie = arg->cookie;
+ res.stat = nlm_granted;
+ res.sequence = 1234356; /* X/Open says this field is ignored? */
+ return (&res);
+}
+
+/* nlm_nm_lock ------------------------------------------------------------ */
+/*
+ Purpose: non-monitored version of nlm_lock()
+ Returns: as for nlm_lock()
+ Notes: These locks are in the same style as the standard nlm_lock,
+ but the rpc.statd should not be called to establish a
+ monitor for the client machine, since that machine is
+ declared not to be running a rpc.statd, and so would not
+ respond to the statd protocol.
+*/
+
+nlm_res *nlm_nm_lock_3_svc(nlm_lockargs *arg, struct svc_req *rqstp)
+{
+ static nlm_res res;
+ if (debug_level) log_from_addr("nlm_nm_lock", rqstp);
+
+ /* copy cookie from arg to result. See comment in nlm_test_1() */
+ res.cookie = arg->cookie;
+ res.stat.stat = nlm_granted;
+ return (&res);
+}
+
+/* nlm_free_all ------------------------------------------------------------ */
+/*
+ Purpose: Release all locks held by a named client
+ Returns: Nothing
+ Notes: Potential denial of service security problem here - the
+ locks to be released are specified by a host name, independent
+ of the address from which the request has arrived.
+ Should probably be rejected if the named host has been
+ using monitored locks.
+*/
+
+void *nlm_free_all_3_svc(nlm_notify *arg, struct svc_req *rqstp)
+{
+ static char dummy;
+
+ if (debug_level) log_from_addr("nlm_free_all", rqstp);
+ return (&dummy);
+}
+
+