2017-06-15 01:37:39 +07:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2016-2017, Mellanox Technologies. All rights reserved.
|
|
|
|
* Copyright (c) 2016-2017, Dave Watson <davejwatson@fb.com>. All rights reserved.
|
|
|
|
*
|
|
|
|
* This software is available to you under a choice of one of two
|
|
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
|
|
* General Public License (GPL) Version 2, available from the file
|
|
|
|
* COPYING in the main directory of this source tree, or the
|
|
|
|
* OpenIB.org BSD license below:
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or
|
|
|
|
* without modification, are permitted provided that the following
|
|
|
|
* conditions are met:
|
|
|
|
*
|
|
|
|
* - Redistributions of source code must retain the above
|
|
|
|
* copyright notice, this list of conditions and the following
|
|
|
|
* disclaimer.
|
|
|
|
*
|
|
|
|
* - Redistributions in binary form must reproduce the above
|
|
|
|
* copyright notice, this list of conditions and the following
|
|
|
|
* disclaimer in the documentation and/or other materials
|
|
|
|
* provided with the distribution.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
|
|
* SOFTWARE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/module.h>
|
|
|
|
|
|
|
|
#include <net/tcp.h>
|
|
|
|
#include <net/inet_common.h>
|
|
|
|
#include <linux/highmem.h>
|
|
|
|
#include <linux/netdevice.h>
|
|
|
|
#include <linux/sched/signal.h>
|
2018-03-31 23:11:52 +07:00
|
|
|
#include <linux/inetdevice.h>
|
2017-06-15 01:37:39 +07:00
|
|
|
|
|
|
|
#include <net/tls.h>
|
|
|
|
|
|
|
|
MODULE_AUTHOR("Mellanox Technologies");
|
|
|
|
MODULE_DESCRIPTION("Transport Layer Security Support");
|
|
|
|
MODULE_LICENSE("Dual BSD/GPL");
|
tcp, ulp: add alias for all ulp modules
Lets not turn the TCP ULP lookup into an arbitrary module loader as
we only intend to load ULP modules through this mechanism, not other
unrelated kernel modules:
[root@bar]# cat foo.c
#include <sys/types.h>
#include <sys/socket.h>
#include <linux/tcp.h>
#include <linux/in.h>
int main(void)
{
int sock = socket(PF_INET, SOCK_STREAM, 0);
setsockopt(sock, IPPROTO_TCP, TCP_ULP, "sctp", sizeof("sctp"));
return 0;
}
[root@bar]# gcc foo.c -O2 -Wall
[root@bar]# lsmod | grep sctp
[root@bar]# ./a.out
[root@bar]# lsmod | grep sctp
sctp 1077248 4
libcrc32c 16384 3 nf_conntrack,nf_nat,sctp
[root@bar]#
Fix it by adding module alias to TCP ULP modules, so probing module
via request_module() will be limited to tcp-ulp-[name]. The existing
modules like kTLS will load fine given tcp-ulp-tls alias, but others
will fail to load:
[root@bar]# lsmod | grep sctp
[root@bar]# ./a.out
[root@bar]# lsmod | grep sctp
[root@bar]#
Sockmap is not affected from this since it's either built-in or not.
Fixes: 734942cc4ea6 ("tcp: ULP infrastructure")
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: John Fastabend <john.fastabend@gmail.com>
Acked-by: Song Liu <songliubraving@fb.com>
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
2018-08-17 02:49:06 +07:00
|
|
|
MODULE_ALIAS_TCP_ULP("tls");
|
2017-06-15 01:37:39 +07:00
|
|
|
|
2018-02-27 19:18:39 +07:00
|
|
|
enum {
|
|
|
|
TLSV4,
|
|
|
|
TLSV6,
|
|
|
|
TLS_NUM_PROTS,
|
|
|
|
};
|
2017-11-13 15:22:45 +07:00
|
|
|
|
2018-02-27 19:18:39 +07:00
|
|
|
static struct proto *saved_tcpv6_prot;
|
|
|
|
static DEFINE_MUTEX(tcpv6_prot_mutex);
|
2018-12-21 02:35:36 +07:00
|
|
|
static struct proto *saved_tcpv4_prot;
|
|
|
|
static DEFINE_MUTEX(tcpv4_prot_mutex);
|
2018-03-31 23:11:52 +07:00
|
|
|
static LIST_HEAD(device_list);
|
2018-12-11 17:20:09 +07:00
|
|
|
static DEFINE_SPINLOCK(device_spinlock);
|
2018-04-30 14:16:15 +07:00
|
|
|
static struct proto tls_prots[TLS_NUM_PROTS][TLS_NUM_CONFIG][TLS_NUM_CONFIG];
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
static struct proto_ops tls_sw_proto_ops;
|
2019-01-18 11:55:53 +07:00
|
|
|
static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG],
|
|
|
|
struct proto *base);
|
2017-11-13 15:22:45 +07:00
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
static void update_sk_prot(struct sock *sk, struct tls_context *ctx)
|
2017-11-13 15:22:45 +07:00
|
|
|
{
|
2018-02-27 19:18:39 +07:00
|
|
|
int ip_ver = sk->sk_family == AF_INET6 ? TLSV6 : TLSV4;
|
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
sk->sk_prot = &tls_prots[ip_ver][ctx->tx_conf][ctx->rx_conf];
|
2017-11-13 15:22:45 +07:00
|
|
|
}
|
2017-06-15 01:37:39 +07:00
|
|
|
|
|
|
|
int wait_on_pending_writer(struct sock *sk, long *timeo)
|
|
|
|
{
|
|
|
|
int rc = 0;
|
|
|
|
DEFINE_WAIT_FUNC(wait, woken_wake_function);
|
|
|
|
|
|
|
|
add_wait_queue(sk_sleep(sk), &wait);
|
|
|
|
while (1) {
|
|
|
|
if (!*timeo) {
|
|
|
|
rc = -EAGAIN;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (signal_pending(current)) {
|
|
|
|
rc = sock_intr_errno(*timeo);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (sk_wait_event(sk, timeo, !sk->sk_write_pending, &wait))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
remove_wait_queue(sk_sleep(sk), &wait);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
int tls_push_sg(struct sock *sk,
|
|
|
|
struct tls_context *ctx,
|
|
|
|
struct scatterlist *sg,
|
|
|
|
u16 first_offset,
|
|
|
|
int flags)
|
|
|
|
{
|
|
|
|
int sendpage_flags = flags | MSG_SENDPAGE_NOTLAST;
|
|
|
|
int ret = 0;
|
|
|
|
struct page *p;
|
|
|
|
size_t size;
|
|
|
|
int offset = first_offset;
|
|
|
|
|
|
|
|
size = sg->length - offset;
|
|
|
|
offset += sg->offset;
|
|
|
|
|
2018-05-02 03:05:39 +07:00
|
|
|
ctx->in_tcp_sendpages = true;
|
2017-06-15 01:37:39 +07:00
|
|
|
while (1) {
|
|
|
|
if (sg_is_last(sg))
|
|
|
|
sendpage_flags = flags;
|
|
|
|
|
|
|
|
/* is sending application-limited? */
|
|
|
|
tcp_rate_check_app_limited(sk);
|
|
|
|
p = sg_page(sg);
|
|
|
|
retry:
|
|
|
|
ret = do_tcp_sendpages(sk, p, offset, size, sendpage_flags);
|
|
|
|
|
|
|
|
if (ret != size) {
|
|
|
|
if (ret > 0) {
|
|
|
|
offset += ret;
|
|
|
|
size -= ret;
|
|
|
|
goto retry;
|
|
|
|
}
|
|
|
|
|
|
|
|
offset -= sg->offset;
|
|
|
|
ctx->partially_sent_offset = offset;
|
|
|
|
ctx->partially_sent_record = (void *)sg;
|
2018-05-07 09:24:39 +07:00
|
|
|
ctx->in_tcp_sendpages = false;
|
2017-06-15 01:37:39 +07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
put_page(p);
|
|
|
|
sk_mem_uncharge(sk, sg->length);
|
|
|
|
sg = sg_next(sg);
|
|
|
|
if (!sg)
|
|
|
|
break;
|
|
|
|
|
|
|
|
offset = sg->offset;
|
|
|
|
size = sg->length;
|
|
|
|
}
|
|
|
|
|
2018-05-02 03:05:39 +07:00
|
|
|
ctx->in_tcp_sendpages = false;
|
2017-06-15 01:37:39 +07:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tls_handle_open_record(struct sock *sk, int flags)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
if (tls_is_pending_open_record(ctx))
|
|
|
|
return ctx->push_pending_record(sk, flags);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int tls_proccess_cmsg(struct sock *sk, struct msghdr *msg,
|
|
|
|
unsigned char *record_type)
|
|
|
|
{
|
|
|
|
struct cmsghdr *cmsg;
|
|
|
|
int rc = -EINVAL;
|
|
|
|
|
|
|
|
for_each_cmsghdr(cmsg, msg) {
|
|
|
|
if (!CMSG_OK(msg, cmsg))
|
|
|
|
return -EINVAL;
|
|
|
|
if (cmsg->cmsg_level != SOL_TLS)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
switch (cmsg->cmsg_type) {
|
|
|
|
case TLS_SET_RECORD_TYPE:
|
|
|
|
if (cmsg->cmsg_len < CMSG_LEN(sizeof(*record_type)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (msg->msg_flags & MSG_MORE)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
rc = tls_handle_open_record(sk, msg->msg_flags);
|
|
|
|
if (rc)
|
|
|
|
return rc;
|
|
|
|
|
|
|
|
*record_type = *(unsigned char *)CMSG_DATA(cmsg);
|
|
|
|
rc = 0;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2018-09-21 11:16:13 +07:00
|
|
|
int tls_push_partial_record(struct sock *sk, struct tls_context *ctx,
|
|
|
|
int flags)
|
2017-06-15 01:37:39 +07:00
|
|
|
{
|
|
|
|
struct scatterlist *sg;
|
|
|
|
u16 offset;
|
|
|
|
|
|
|
|
sg = ctx->partially_sent_record;
|
|
|
|
offset = ctx->partially_sent_offset;
|
|
|
|
|
|
|
|
ctx->partially_sent_record = NULL;
|
|
|
|
return tls_push_sg(sk, ctx, sg, offset, flags);
|
|
|
|
}
|
|
|
|
|
2019-04-11 01:04:31 +07:00
|
|
|
bool tls_free_partial_record(struct sock *sk, struct tls_context *ctx)
|
|
|
|
{
|
|
|
|
struct scatterlist *sg;
|
|
|
|
|
|
|
|
sg = ctx->partially_sent_record;
|
|
|
|
if (!sg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
while (1) {
|
|
|
|
put_page(sg_page(sg));
|
|
|
|
sk_mem_uncharge(sk, sg->length);
|
|
|
|
|
|
|
|
if (sg_is_last(sg))
|
|
|
|
break;
|
|
|
|
sg++;
|
|
|
|
}
|
|
|
|
ctx->partially_sent_record = NULL;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2017-06-15 01:37:39 +07:00
|
|
|
static void tls_write_space(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
2018-08-22 22:37:32 +07:00
|
|
|
/* If in_tcp_sendpages call lower protocol write space handler
|
|
|
|
* to ensure we wake up any waiting operations there. For example
|
|
|
|
* if do_tcp_sendpages where to call sk_wait_event.
|
|
|
|
*/
|
|
|
|
if (ctx->in_tcp_sendpages) {
|
|
|
|
ctx->sk_write_space(sk);
|
2018-05-02 03:05:39 +07:00
|
|
|
return;
|
2018-08-22 22:37:32 +07:00
|
|
|
}
|
2018-05-02 03:05:39 +07:00
|
|
|
|
2019-02-27 22:38:04 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
if (ctx->tx_conf == TLS_HW)
|
|
|
|
tls_device_write_space(sk, ctx);
|
|
|
|
else
|
|
|
|
#endif
|
|
|
|
tls_sw_write_space(sk, ctx);
|
2019-03-12 15:22:57 +07:00
|
|
|
|
|
|
|
ctx->sk_write_space(sk);
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
2019-06-29 06:11:39 +07:00
|
|
|
void tls_ctx_free(struct tls_context *ctx)
|
2018-09-12 22:44:42 +07:00
|
|
|
{
|
|
|
|
if (!ctx)
|
|
|
|
return;
|
|
|
|
|
|
|
|
memzero_explicit(&ctx->crypto_send, sizeof(ctx->crypto_send));
|
|
|
|
memzero_explicit(&ctx->crypto_recv, sizeof(ctx->crypto_recv));
|
|
|
|
kfree(ctx);
|
|
|
|
}
|
|
|
|
|
2019-07-20 00:29:18 +07:00
|
|
|
static void tls_ctx_free_deferred(struct work_struct *gc)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = container_of(gc, struct tls_context, gc);
|
|
|
|
|
|
|
|
/* Ensure any remaining work items are completed. The sk will
|
|
|
|
* already have lost its tls_ctx reference by the time we get
|
|
|
|
* here so no xmit operation will actually be performed.
|
|
|
|
*/
|
|
|
|
if (ctx->tx_conf == TLS_SW) {
|
|
|
|
tls_sw_cancel_work_tx(ctx);
|
|
|
|
tls_sw_free_ctx_tx(ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ctx->rx_conf == TLS_SW) {
|
|
|
|
tls_sw_strparser_done(ctx);
|
|
|
|
tls_sw_free_ctx_rx(ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
tls_ctx_free(ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tls_ctx_free_wq(struct tls_context *ctx)
|
|
|
|
{
|
|
|
|
INIT_WORK(&ctx->gc, tls_ctx_free_deferred);
|
|
|
|
schedule_work(&ctx->gc);
|
|
|
|
}
|
|
|
|
|
2019-07-20 00:29:17 +07:00
|
|
|
static void tls_sk_proto_cleanup(struct sock *sk,
|
|
|
|
struct tls_context *ctx, long timeo)
|
2017-06-15 01:37:39 +07:00
|
|
|
{
|
2019-06-24 11:26:58 +07:00
|
|
|
if (unlikely(sk->sk_write_pending) &&
|
|
|
|
!wait_on_pending_writer(sk, &timeo))
|
2017-06-15 01:37:39 +07:00
|
|
|
tls_handle_open_record(sk, 0);
|
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
/* We need these for tls_sw_fallback handling of other packets */
|
|
|
|
if (ctx->tx_conf == TLS_SW) {
|
|
|
|
kfree(ctx->tx.rec_seq);
|
|
|
|
kfree(ctx->tx.iv);
|
2019-07-20 00:29:17 +07:00
|
|
|
tls_sw_release_resources_tx(sk);
|
2019-04-11 06:23:39 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
2019-04-11 01:04:31 +07:00
|
|
|
} else if (ctx->tx_conf == TLS_HW) {
|
|
|
|
tls_device_free_resources_tx(sk);
|
2019-04-11 06:23:39 +07:00
|
|
|
#endif
|
2018-04-30 14:16:15 +07:00
|
|
|
}
|
2017-06-15 01:37:39 +07:00
|
|
|
|
2019-04-20 06:52:19 +07:00
|
|
|
if (ctx->rx_conf == TLS_SW)
|
2019-07-20 00:29:17 +07:00
|
|
|
tls_sw_release_resources_rx(sk);
|
2017-06-15 01:37:39 +07:00
|
|
|
|
2018-04-30 14:16:16 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
2018-07-13 18:33:43 +07:00
|
|
|
if (ctx->rx_conf == TLS_HW)
|
|
|
|
tls_device_offload_cleanup_rx(sk);
|
2018-04-30 14:16:16 +07:00
|
|
|
#endif
|
2019-07-20 00:29:17 +07:00
|
|
|
}
|
|
|
|
|
2019-07-20 00:29:18 +07:00
|
|
|
static void tls_sk_proto_unhash(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct inet_connection_sock *icsk = inet_csk(sk);
|
|
|
|
long timeo = sock_sndtimeo(sk, 0);
|
|
|
|
struct tls_context *ctx;
|
|
|
|
|
|
|
|
if (unlikely(!icsk->icsk_ulp_data)) {
|
|
|
|
if (sk->sk_prot->unhash)
|
|
|
|
sk->sk_prot->unhash(sk);
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx = tls_get_ctx(sk);
|
|
|
|
tls_sk_proto_cleanup(sk, ctx, timeo);
|
2019-07-20 00:29:22 +07:00
|
|
|
write_lock_bh(&sk->sk_callback_lock);
|
2019-07-20 00:29:18 +07:00
|
|
|
icsk->icsk_ulp_data = NULL;
|
2019-07-20 00:29:22 +07:00
|
|
|
sk->sk_prot = ctx->sk_proto;
|
|
|
|
write_unlock_bh(&sk->sk_callback_lock);
|
2019-07-20 00:29:18 +07:00
|
|
|
|
|
|
|
if (ctx->sk_proto->unhash)
|
|
|
|
ctx->sk_proto->unhash(sk);
|
|
|
|
tls_ctx_free_wq(ctx);
|
|
|
|
}
|
|
|
|
|
2019-07-20 00:29:17 +07:00
|
|
|
static void tls_sk_proto_close(struct sock *sk, long timeout)
|
|
|
|
{
|
2019-07-20 00:29:22 +07:00
|
|
|
struct inet_connection_sock *icsk = inet_csk(sk);
|
2019-07-20 00:29:17 +07:00
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
long timeo = sock_sndtimeo(sk, 0);
|
|
|
|
bool free_ctx;
|
|
|
|
|
|
|
|
if (ctx->tx_conf == TLS_SW)
|
|
|
|
tls_sw_cancel_work_tx(ctx);
|
|
|
|
|
|
|
|
lock_sock(sk);
|
|
|
|
free_ctx = ctx->tx_conf != TLS_HW && ctx->rx_conf != TLS_HW;
|
|
|
|
|
|
|
|
if (ctx->tx_conf != TLS_BASE || ctx->rx_conf != TLS_BASE)
|
|
|
|
tls_sk_proto_cleanup(sk, ctx, timeo);
|
2018-04-30 14:16:16 +07:00
|
|
|
|
2019-07-20 00:29:22 +07:00
|
|
|
write_lock_bh(&sk->sk_callback_lock);
|
|
|
|
if (free_ctx)
|
|
|
|
icsk->icsk_ulp_data = NULL;
|
2019-07-20 00:29:18 +07:00
|
|
|
sk->sk_prot = ctx->sk_proto;
|
2019-07-20 00:29:22 +07:00
|
|
|
write_unlock_bh(&sk->sk_callback_lock);
|
2017-06-15 01:37:39 +07:00
|
|
|
release_sock(sk);
|
2019-07-20 00:29:17 +07:00
|
|
|
if (ctx->tx_conf == TLS_SW)
|
|
|
|
tls_sw_free_ctx_tx(ctx);
|
|
|
|
if (ctx->rx_conf == TLS_SW || ctx->rx_conf == TLS_HW)
|
|
|
|
tls_sw_strparser_done(ctx);
|
|
|
|
if (ctx->rx_conf == TLS_SW)
|
|
|
|
tls_sw_free_ctx_rx(ctx);
|
2019-07-20 00:29:22 +07:00
|
|
|
ctx->sk_proto_close(sk, timeout);
|
2019-07-20 00:29:17 +07:00
|
|
|
|
2018-05-05 22:35:04 +07:00
|
|
|
if (free_ctx)
|
2018-09-12 22:44:42 +07:00
|
|
|
tls_ctx_free(ctx);
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int do_tls_getsockopt_tx(struct sock *sk, char __user *optval,
|
|
|
|
int __user *optlen)
|
|
|
|
{
|
|
|
|
int rc = 0;
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_crypto_info *crypto_info;
|
|
|
|
int len;
|
|
|
|
|
|
|
|
if (get_user(len, optlen))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
if (!optval || (len < sizeof(*crypto_info))) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!ctx) {
|
|
|
|
rc = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* get user crypto info */
|
2018-09-12 22:44:42 +07:00
|
|
|
crypto_info = &ctx->crypto_send.info;
|
2017-06-15 01:37:39 +07:00
|
|
|
|
|
|
|
if (!TLS_CRYPTO_INFO_READY(crypto_info)) {
|
|
|
|
rc = -EBUSY;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2017-07-06 11:56:36 +07:00
|
|
|
if (len == sizeof(*crypto_info)) {
|
2017-06-23 17:15:44 +07:00
|
|
|
if (copy_to_user(optval, crypto_info, sizeof(*crypto_info)))
|
|
|
|
rc = -EFAULT;
|
2017-06-15 01:37:39 +07:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (crypto_info->cipher_type) {
|
|
|
|
case TLS_CIPHER_AES_GCM_128: {
|
|
|
|
struct tls12_crypto_info_aes_gcm_128 *
|
|
|
|
crypto_info_aes_gcm_128 =
|
|
|
|
container_of(crypto_info,
|
|
|
|
struct tls12_crypto_info_aes_gcm_128,
|
|
|
|
info);
|
|
|
|
|
|
|
|
if (len != sizeof(*crypto_info_aes_gcm_128)) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
lock_sock(sk);
|
2018-02-14 15:46:06 +07:00
|
|
|
memcpy(crypto_info_aes_gcm_128->iv,
|
2018-03-23 00:10:06 +07:00
|
|
|
ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE,
|
2017-06-15 01:37:39 +07:00
|
|
|
TLS_CIPHER_AES_GCM_128_IV_SIZE);
|
2018-03-23 00:10:06 +07:00
|
|
|
memcpy(crypto_info_aes_gcm_128->rec_seq, ctx->tx.rec_seq,
|
2018-02-14 15:46:08 +07:00
|
|
|
TLS_CIPHER_AES_GCM_128_REC_SEQ_SIZE);
|
2017-06-15 01:37:39 +07:00
|
|
|
release_sock(sk);
|
2017-06-23 17:15:44 +07:00
|
|
|
if (copy_to_user(optval,
|
|
|
|
crypto_info_aes_gcm_128,
|
|
|
|
sizeof(*crypto_info_aes_gcm_128)))
|
|
|
|
rc = -EFAULT;
|
2017-06-15 01:37:39 +07:00
|
|
|
break;
|
|
|
|
}
|
2019-01-31 04:58:05 +07:00
|
|
|
case TLS_CIPHER_AES_GCM_256: {
|
|
|
|
struct tls12_crypto_info_aes_gcm_256 *
|
|
|
|
crypto_info_aes_gcm_256 =
|
|
|
|
container_of(crypto_info,
|
|
|
|
struct tls12_crypto_info_aes_gcm_256,
|
|
|
|
info);
|
|
|
|
|
|
|
|
if (len != sizeof(*crypto_info_aes_gcm_256)) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
lock_sock(sk);
|
|
|
|
memcpy(crypto_info_aes_gcm_256->iv,
|
|
|
|
ctx->tx.iv + TLS_CIPHER_AES_GCM_256_SALT_SIZE,
|
|
|
|
TLS_CIPHER_AES_GCM_256_IV_SIZE);
|
|
|
|
memcpy(crypto_info_aes_gcm_256->rec_seq, ctx->tx.rec_seq,
|
|
|
|
TLS_CIPHER_AES_GCM_256_REC_SEQ_SIZE);
|
|
|
|
release_sock(sk);
|
|
|
|
if (copy_to_user(optval,
|
|
|
|
crypto_info_aes_gcm_256,
|
|
|
|
sizeof(*crypto_info_aes_gcm_256)))
|
|
|
|
rc = -EFAULT;
|
|
|
|
break;
|
|
|
|
}
|
2017-06-15 01:37:39 +07:00
|
|
|
default:
|
|
|
|
rc = -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int do_tls_getsockopt(struct sock *sk, int optname,
|
|
|
|
char __user *optval, int __user *optlen)
|
|
|
|
{
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
switch (optname) {
|
|
|
|
case TLS_TX:
|
|
|
|
rc = do_tls_getsockopt_tx(sk, optval, optlen);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
rc = -ENOPROTOOPT;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tls_getsockopt(struct sock *sk, int level, int optname,
|
|
|
|
char __user *optval, int __user *optlen)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
if (level != SOL_TLS)
|
|
|
|
return ctx->getsockopt(sk, level, optname, optval, optlen);
|
|
|
|
|
|
|
|
return do_tls_getsockopt(sk, optname, optval, optlen);
|
|
|
|
}
|
|
|
|
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
static int do_tls_setsockopt_conf(struct sock *sk, char __user *optval,
|
|
|
|
unsigned int optlen, int tx)
|
2017-06-15 01:37:39 +07:00
|
|
|
{
|
2017-11-13 15:22:48 +07:00
|
|
|
struct tls_crypto_info *crypto_info;
|
2019-02-14 14:11:35 +07:00
|
|
|
struct tls_crypto_info *alt_crypto_info;
|
2017-06-15 01:37:39 +07:00
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
2019-01-31 04:58:05 +07:00
|
|
|
size_t optsize;
|
2017-06-15 01:37:39 +07:00
|
|
|
int rc = 0;
|
2018-03-23 00:10:26 +07:00
|
|
|
int conf;
|
2017-06-15 01:37:39 +07:00
|
|
|
|
|
|
|
if (!optval || (optlen < sizeof(*crypto_info))) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2019-02-14 14:11:35 +07:00
|
|
|
if (tx) {
|
2018-09-12 22:44:42 +07:00
|
|
|
crypto_info = &ctx->crypto_send.info;
|
2019-02-14 14:11:35 +07:00
|
|
|
alt_crypto_info = &ctx->crypto_recv.info;
|
|
|
|
} else {
|
2018-09-12 22:44:42 +07:00
|
|
|
crypto_info = &ctx->crypto_recv.info;
|
2019-02-14 14:11:35 +07:00
|
|
|
alt_crypto_info = &ctx->crypto_send.info;
|
|
|
|
}
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
|
2017-11-13 15:22:48 +07:00
|
|
|
/* Currently we don't support set crypto info more than one time */
|
2018-01-16 22:04:27 +07:00
|
|
|
if (TLS_CRYPTO_INFO_READY(crypto_info)) {
|
|
|
|
rc = -EBUSY;
|
2017-11-13 15:22:48 +07:00
|
|
|
goto out;
|
2018-01-16 22:04:27 +07:00
|
|
|
}
|
2017-11-13 15:22:48 +07:00
|
|
|
|
|
|
|
rc = copy_from_user(crypto_info, optval, sizeof(*crypto_info));
|
2017-06-15 01:37:39 +07:00
|
|
|
if (rc) {
|
|
|
|
rc = -EFAULT;
|
2018-02-14 15:46:07 +07:00
|
|
|
goto err_crypto_info;
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/* check version */
|
2019-01-31 04:58:31 +07:00
|
|
|
if (crypto_info->version != TLS_1_2_VERSION &&
|
|
|
|
crypto_info->version != TLS_1_3_VERSION) {
|
2017-06-15 01:37:39 +07:00
|
|
|
rc = -ENOTSUPP;
|
2017-11-13 15:22:48 +07:00
|
|
|
goto err_crypto_info;
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
2019-02-14 14:11:35 +07:00
|
|
|
/* Ensure that TLS version and ciphers are same in both directions */
|
|
|
|
if (TLS_CRYPTO_INFO_READY(alt_crypto_info)) {
|
|
|
|
if (alt_crypto_info->version != crypto_info->version ||
|
|
|
|
alt_crypto_info->cipher_type != crypto_info->cipher_type) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto err_crypto_info;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-11-13 15:22:48 +07:00
|
|
|
switch (crypto_info->cipher_type) {
|
2019-01-31 04:58:05 +07:00
|
|
|
case TLS_CIPHER_AES_GCM_128:
|
2019-03-20 09:03:36 +07:00
|
|
|
optsize = sizeof(struct tls12_crypto_info_aes_gcm_128);
|
|
|
|
break;
|
2019-01-31 04:58:05 +07:00
|
|
|
case TLS_CIPHER_AES_GCM_256: {
|
2019-03-20 09:03:36 +07:00
|
|
|
optsize = sizeof(struct tls12_crypto_info_aes_gcm_256);
|
2017-06-15 01:37:39 +07:00
|
|
|
break;
|
|
|
|
}
|
2019-03-20 09:03:36 +07:00
|
|
|
case TLS_CIPHER_AES_CCM_128:
|
|
|
|
optsize = sizeof(struct tls12_crypto_info_aes_ccm_128);
|
|
|
|
break;
|
2017-06-15 01:37:39 +07:00
|
|
|
default:
|
|
|
|
rc = -EINVAL;
|
tls: reset crypto_info when do_tls_setsockopt_tx fails
The current code copies directly from userspace to ctx->crypto_send, but
doesn't always reinitialize it to 0 on failure. This causes any
subsequent attempt to use this setsockopt to fail because of the
TLS_CRYPTO_INFO_READY check, eventhough crypto_info is not actually
ready.
This should result in a correctly set up socket after the 3rd call, but
currently it does not:
size_t s = sizeof(struct tls12_crypto_info_aes_gcm_128);
struct tls12_crypto_info_aes_gcm_128 crypto_good = {
.info.version = TLS_1_2_VERSION,
.info.cipher_type = TLS_CIPHER_AES_GCM_128,
};
struct tls12_crypto_info_aes_gcm_128 crypto_bad_type = crypto_good;
crypto_bad_type.info.cipher_type = 42;
setsockopt(sock, SOL_TLS, TLS_TX, &crypto_bad_type, s);
setsockopt(sock, SOL_TLS, TLS_TX, &crypto_good, s - 1);
setsockopt(sock, SOL_TLS, TLS_TX, &crypto_good, s);
Fixes: 3c4d7559159b ("tls: kernel TLS support")
Signed-off-by: Sabrina Dubroca <sd@queasysnail.net>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-01-16 22:04:28 +07:00
|
|
|
goto err_crypto_info;
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
2019-03-20 09:03:36 +07:00
|
|
|
if (optlen != optsize) {
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto err_crypto_info;
|
|
|
|
}
|
|
|
|
|
|
|
|
rc = copy_from_user(crypto_info + 1, optval + sizeof(*crypto_info),
|
|
|
|
optlen - sizeof(*crypto_info));
|
|
|
|
if (rc) {
|
|
|
|
rc = -EFAULT;
|
|
|
|
goto err_crypto_info;
|
|
|
|
}
|
|
|
|
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
if (tx) {
|
2018-04-30 14:16:16 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
rc = tls_set_device_offload(sk, ctx);
|
|
|
|
conf = TLS_HW;
|
|
|
|
if (rc) {
|
|
|
|
#else
|
|
|
|
{
|
|
|
|
#endif
|
|
|
|
rc = tls_set_sw_offload(sk, ctx, 1);
|
2019-07-20 00:29:14 +07:00
|
|
|
if (rc)
|
|
|
|
goto err_crypto_info;
|
2018-04-30 14:16:16 +07:00
|
|
|
conf = TLS_SW;
|
|
|
|
}
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
} else {
|
2018-07-13 18:33:43 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
rc = tls_set_device_offload_rx(sk, ctx);
|
|
|
|
conf = TLS_HW;
|
|
|
|
if (rc) {
|
|
|
|
#else
|
|
|
|
{
|
|
|
|
#endif
|
|
|
|
rc = tls_set_sw_offload(sk, ctx, 0);
|
2019-07-20 00:29:14 +07:00
|
|
|
if (rc)
|
|
|
|
goto err_crypto_info;
|
2018-07-13 18:33:43 +07:00
|
|
|
conf = TLS_SW;
|
|
|
|
}
|
2019-07-20 00:29:17 +07:00
|
|
|
tls_sw_strparser_arm(sk, ctx);
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
}
|
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
if (tx)
|
|
|
|
ctx->tx_conf = conf;
|
|
|
|
else
|
|
|
|
ctx->rx_conf = conf;
|
2017-11-13 15:22:45 +07:00
|
|
|
update_sk_prot(sk, ctx);
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
if (tx) {
|
|
|
|
ctx->sk_write_space = sk->sk_write_space;
|
|
|
|
sk->sk_write_space = tls_write_space;
|
|
|
|
} else {
|
|
|
|
sk->sk_socket->ops = &tls_sw_proto_ops;
|
|
|
|
}
|
2017-06-15 01:37:39 +07:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
err_crypto_info:
|
2018-09-12 22:44:43 +07:00
|
|
|
memzero_explicit(crypto_info, sizeof(union tls_crypto_context));
|
2017-06-15 01:37:39 +07:00
|
|
|
out:
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int do_tls_setsockopt(struct sock *sk, int optname,
|
|
|
|
char __user *optval, unsigned int optlen)
|
|
|
|
{
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
switch (optname) {
|
|
|
|
case TLS_TX:
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
case TLS_RX:
|
2017-06-15 01:37:39 +07:00
|
|
|
lock_sock(sk);
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
rc = do_tls_setsockopt_conf(sk, optval, optlen,
|
|
|
|
optname == TLS_TX);
|
2017-06-15 01:37:39 +07:00
|
|
|
release_sock(sk);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
rc = -ENOPROTOOPT;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tls_setsockopt(struct sock *sk, int level, int optname,
|
|
|
|
char __user *optval, unsigned int optlen)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
|
|
|
|
if (level != SOL_TLS)
|
|
|
|
return ctx->setsockopt(sk, level, optname, optval, optlen);
|
|
|
|
|
|
|
|
return do_tls_setsockopt(sk, optname, optval, optlen);
|
|
|
|
}
|
|
|
|
|
2018-03-31 23:11:52 +07:00
|
|
|
static struct tls_context *create_ctx(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct inet_connection_sock *icsk = inet_csk(sk);
|
|
|
|
struct tls_context *ctx;
|
|
|
|
|
2018-12-19 18:48:22 +07:00
|
|
|
ctx = kzalloc(sizeof(*ctx), GFP_ATOMIC);
|
2018-03-31 23:11:52 +07:00
|
|
|
if (!ctx)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
icsk->icsk_ulp_data = ctx;
|
2018-12-11 17:19:40 +07:00
|
|
|
ctx->setsockopt = sk->sk_prot->setsockopt;
|
|
|
|
ctx->getsockopt = sk->sk_prot->getsockopt;
|
|
|
|
ctx->sk_proto_close = sk->sk_prot->close;
|
2019-07-20 00:29:18 +07:00
|
|
|
ctx->unhash = sk->sk_prot->unhash;
|
2018-03-31 23:11:52 +07:00
|
|
|
return ctx;
|
|
|
|
}
|
|
|
|
|
2019-01-18 11:55:53 +07:00
|
|
|
static void tls_build_proto(struct sock *sk)
|
|
|
|
{
|
|
|
|
int ip_ver = sk->sk_family == AF_INET6 ? TLSV6 : TLSV4;
|
|
|
|
|
|
|
|
/* Build IPv6 TLS whenever the address of tcpv6 _prot changes */
|
|
|
|
if (ip_ver == TLSV6 &&
|
|
|
|
unlikely(sk->sk_prot != smp_load_acquire(&saved_tcpv6_prot))) {
|
|
|
|
mutex_lock(&tcpv6_prot_mutex);
|
|
|
|
if (likely(sk->sk_prot != saved_tcpv6_prot)) {
|
|
|
|
build_protos(tls_prots[TLSV6], sk->sk_prot);
|
|
|
|
smp_store_release(&saved_tcpv6_prot, sk->sk_prot);
|
|
|
|
}
|
|
|
|
mutex_unlock(&tcpv6_prot_mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ip_ver == TLSV4 &&
|
|
|
|
unlikely(sk->sk_prot != smp_load_acquire(&saved_tcpv4_prot))) {
|
|
|
|
mutex_lock(&tcpv4_prot_mutex);
|
|
|
|
if (likely(sk->sk_prot != saved_tcpv4_prot)) {
|
|
|
|
build_protos(tls_prots[TLSV4], sk->sk_prot);
|
|
|
|
smp_store_release(&saved_tcpv4_prot, sk->sk_prot);
|
|
|
|
}
|
|
|
|
mutex_unlock(&tcpv4_prot_mutex);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-18 11:56:21 +07:00
|
|
|
static void tls_hw_sk_destruct(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
struct inet_connection_sock *icsk = inet_csk(sk);
|
|
|
|
|
|
|
|
ctx->sk_destruct(sk);
|
|
|
|
/* Free ctx */
|
2019-06-29 06:11:39 +07:00
|
|
|
tls_ctx_free(ctx);
|
2019-01-18 11:56:21 +07:00
|
|
|
icsk->icsk_ulp_data = NULL;
|
|
|
|
}
|
|
|
|
|
2018-03-31 23:11:52 +07:00
|
|
|
static int tls_hw_prot(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx;
|
|
|
|
struct tls_device *dev;
|
|
|
|
int rc = 0;
|
|
|
|
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
list_for_each_entry(dev, &device_list, dev_list) {
|
|
|
|
if (dev->feature && dev->feature(dev)) {
|
|
|
|
ctx = create_ctx(sk);
|
|
|
|
if (!ctx)
|
|
|
|
goto out;
|
|
|
|
|
2019-01-18 11:55:53 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
|
|
|
tls_build_proto(sk);
|
2018-03-31 23:11:52 +07:00
|
|
|
ctx->hash = sk->sk_prot->hash;
|
|
|
|
ctx->unhash = sk->sk_prot->unhash;
|
|
|
|
ctx->sk_proto_close = sk->sk_prot->close;
|
2019-01-18 11:56:21 +07:00
|
|
|
ctx->sk_destruct = sk->sk_destruct;
|
|
|
|
sk->sk_destruct = tls_hw_sk_destruct;
|
2018-04-30 14:16:15 +07:00
|
|
|
ctx->rx_conf = TLS_HW_RECORD;
|
|
|
|
ctx->tx_conf = TLS_HW_RECORD;
|
2018-03-31 23:11:52 +07:00
|
|
|
update_sk_prot(sk, ctx);
|
2019-01-18 11:55:53 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
rc = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
out:
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tls_hw_unhash(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_device *dev;
|
|
|
|
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
list_for_each_entry(dev, &device_list, dev_list) {
|
2018-12-11 17:20:09 +07:00
|
|
|
if (dev->unhash) {
|
|
|
|
kref_get(&dev->kref);
|
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
dev->unhash(dev, sk);
|
2018-12-11 17:20:09 +07:00
|
|
|
kref_put(&dev->kref, dev->release);
|
|
|
|
spin_lock_bh(&device_spinlock);
|
|
|
|
}
|
2018-03-31 23:11:52 +07:00
|
|
|
}
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
ctx->unhash(sk);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tls_hw_hash(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx = tls_get_ctx(sk);
|
|
|
|
struct tls_device *dev;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
err = ctx->hash(sk);
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
list_for_each_entry(dev, &device_list, dev_list) {
|
2018-12-11 17:20:09 +07:00
|
|
|
if (dev->hash) {
|
|
|
|
kref_get(&dev->kref);
|
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
err |= dev->hash(dev, sk);
|
2018-12-11 17:20:09 +07:00
|
|
|
kref_put(&dev->kref, dev->release);
|
|
|
|
spin_lock_bh(&device_spinlock);
|
|
|
|
}
|
2018-03-31 23:11:52 +07:00
|
|
|
}
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
|
|
|
|
if (err)
|
|
|
|
tls_hw_unhash(sk);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG],
|
|
|
|
struct proto *base)
|
2018-02-27 19:18:39 +07:00
|
|
|
{
|
2018-04-30 14:16:15 +07:00
|
|
|
prot[TLS_BASE][TLS_BASE] = *base;
|
|
|
|
prot[TLS_BASE][TLS_BASE].setsockopt = tls_setsockopt;
|
|
|
|
prot[TLS_BASE][TLS_BASE].getsockopt = tls_getsockopt;
|
|
|
|
prot[TLS_BASE][TLS_BASE].close = tls_sk_proto_close;
|
2019-07-20 00:29:18 +07:00
|
|
|
prot[TLS_BASE][TLS_BASE].unhash = tls_sk_proto_unhash;
|
2018-04-30 14:16:15 +07:00
|
|
|
|
|
|
|
prot[TLS_SW][TLS_BASE] = prot[TLS_BASE][TLS_BASE];
|
|
|
|
prot[TLS_SW][TLS_BASE].sendmsg = tls_sw_sendmsg;
|
|
|
|
prot[TLS_SW][TLS_BASE].sendpage = tls_sw_sendpage;
|
|
|
|
|
|
|
|
prot[TLS_BASE][TLS_SW] = prot[TLS_BASE][TLS_BASE];
|
2018-10-13 07:46:00 +07:00
|
|
|
prot[TLS_BASE][TLS_SW].recvmsg = tls_sw_recvmsg;
|
|
|
|
prot[TLS_BASE][TLS_SW].stream_memory_read = tls_sw_stream_read;
|
|
|
|
prot[TLS_BASE][TLS_SW].close = tls_sk_proto_close;
|
2018-04-30 14:16:15 +07:00
|
|
|
|
|
|
|
prot[TLS_SW][TLS_SW] = prot[TLS_SW][TLS_BASE];
|
2018-10-13 07:46:00 +07:00
|
|
|
prot[TLS_SW][TLS_SW].recvmsg = tls_sw_recvmsg;
|
|
|
|
prot[TLS_SW][TLS_SW].stream_memory_read = tls_sw_stream_read;
|
|
|
|
prot[TLS_SW][TLS_SW].close = tls_sk_proto_close;
|
2018-04-30 14:16:15 +07:00
|
|
|
|
2018-04-30 14:16:16 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
prot[TLS_HW][TLS_BASE] = prot[TLS_BASE][TLS_BASE];
|
2019-07-20 00:29:18 +07:00
|
|
|
prot[TLS_HW][TLS_BASE].unhash = base->unhash;
|
2018-04-30 14:16:16 +07:00
|
|
|
prot[TLS_HW][TLS_BASE].sendmsg = tls_device_sendmsg;
|
|
|
|
prot[TLS_HW][TLS_BASE].sendpage = tls_device_sendpage;
|
|
|
|
|
|
|
|
prot[TLS_HW][TLS_SW] = prot[TLS_BASE][TLS_SW];
|
2019-07-20 00:29:18 +07:00
|
|
|
prot[TLS_HW][TLS_SW].unhash = base->unhash;
|
2018-04-30 14:16:16 +07:00
|
|
|
prot[TLS_HW][TLS_SW].sendmsg = tls_device_sendmsg;
|
|
|
|
prot[TLS_HW][TLS_SW].sendpage = tls_device_sendpage;
|
2018-07-13 18:33:43 +07:00
|
|
|
|
|
|
|
prot[TLS_BASE][TLS_HW] = prot[TLS_BASE][TLS_SW];
|
2019-07-20 00:29:18 +07:00
|
|
|
prot[TLS_BASE][TLS_HW].unhash = base->unhash;
|
2018-07-13 18:33:43 +07:00
|
|
|
|
|
|
|
prot[TLS_SW][TLS_HW] = prot[TLS_SW][TLS_SW];
|
2019-07-20 00:29:18 +07:00
|
|
|
prot[TLS_SW][TLS_HW].unhash = base->unhash;
|
2018-07-13 18:33:43 +07:00
|
|
|
|
|
|
|
prot[TLS_HW][TLS_HW] = prot[TLS_HW][TLS_SW];
|
2018-04-30 14:16:16 +07:00
|
|
|
#endif
|
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
prot[TLS_HW_RECORD][TLS_HW_RECORD] = *base;
|
|
|
|
prot[TLS_HW_RECORD][TLS_HW_RECORD].hash = tls_hw_hash;
|
|
|
|
prot[TLS_HW_RECORD][TLS_HW_RECORD].unhash = tls_hw_unhash;
|
2018-02-27 19:18:39 +07:00
|
|
|
}
|
|
|
|
|
2017-06-15 01:37:39 +07:00
|
|
|
static int tls_init(struct sock *sk)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx;
|
|
|
|
int rc = 0;
|
|
|
|
|
2018-03-31 23:11:52 +07:00
|
|
|
if (tls_hw_prot(sk))
|
2019-07-20 00:29:22 +07:00
|
|
|
return 0;
|
2018-03-31 23:11:52 +07:00
|
|
|
|
2018-01-16 20:31:52 +07:00
|
|
|
/* The TLS ulp is currently supported only for TCP sockets
|
|
|
|
* in ESTABLISHED state.
|
|
|
|
* Supporting sockets in LISTEN state will require us
|
|
|
|
* to modify the accept implementation to clone rather then
|
|
|
|
* share the ulp context.
|
|
|
|
*/
|
|
|
|
if (sk->sk_state != TCP_ESTABLISHED)
|
|
|
|
return -ENOTSUPP;
|
|
|
|
|
2019-07-20 00:29:22 +07:00
|
|
|
tls_build_proto(sk);
|
|
|
|
|
2017-06-15 01:37:39 +07:00
|
|
|
/* allocate tls context */
|
2019-07-20 00:29:22 +07:00
|
|
|
write_lock_bh(&sk->sk_callback_lock);
|
2018-03-31 23:11:52 +07:00
|
|
|
ctx = create_ctx(sk);
|
2017-06-15 01:37:39 +07:00
|
|
|
if (!ctx) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto out;
|
|
|
|
}
|
2017-11-13 15:22:45 +07:00
|
|
|
|
2018-04-30 14:16:15 +07:00
|
|
|
ctx->tx_conf = TLS_BASE;
|
|
|
|
ctx->rx_conf = TLS_BASE;
|
2019-07-20 00:29:18 +07:00
|
|
|
ctx->sk_proto = sk->sk_prot;
|
2017-11-13 15:22:45 +07:00
|
|
|
update_sk_prot(sk, ctx);
|
2017-06-15 01:37:39 +07:00
|
|
|
out:
|
2019-07-20 00:29:22 +07:00
|
|
|
write_unlock_bh(&sk->sk_callback_lock);
|
2017-06-15 01:37:39 +07:00
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2019-07-20 00:29:22 +07:00
|
|
|
static void tls_update(struct sock *sk, struct proto *p)
|
|
|
|
{
|
|
|
|
struct tls_context *ctx;
|
|
|
|
|
|
|
|
ctx = tls_get_ctx(sk);
|
|
|
|
if (likely(ctx)) {
|
|
|
|
ctx->sk_proto_close = p->close;
|
|
|
|
ctx->sk_proto = p;
|
|
|
|
} else {
|
|
|
|
sk->sk_prot = p;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-31 23:11:52 +07:00
|
|
|
void tls_register_device(struct tls_device *device)
|
|
|
|
{
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
list_add_tail(&device->dev_list, &device_list);
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tls_register_device);
|
|
|
|
|
|
|
|
void tls_unregister_device(struct tls_device *device)
|
|
|
|
{
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_lock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
list_del(&device->dev_list);
|
2018-12-11 17:20:09 +07:00
|
|
|
spin_unlock_bh(&device_spinlock);
|
2018-03-31 23:11:52 +07:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tls_unregister_device);
|
|
|
|
|
2017-06-15 01:37:39 +07:00
|
|
|
static struct tcp_ulp_ops tcp_tls_ulp_ops __read_mostly = {
|
|
|
|
.name = "tls",
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
.init = tls_init,
|
2019-07-20 00:29:22 +07:00
|
|
|
.update = tls_update,
|
2017-06-15 01:37:39 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
static int __init tls_register(void)
|
|
|
|
{
|
tls: RX path for ktls
Add rx path for tls software implementation.
recvmsg, splice_read, and poll implemented.
An additional sockopt TLS_RX is added, with the same interface as
TLS_TX. Either TLX_RX or TLX_TX may be provided separately, or
together (with two different setsockopt calls with appropriate keys).
Control messages are passed via CMSG in a similar way to transmit.
If no cmsg buffer is passed, then only application data records
will be passed to userspace, and EIO is returned for other types of
alerts.
EBADMSG is passed for decryption errors, and EMSGSIZE is passed for
framing too big, and EBADMSG for framing too small (matching openssl
semantics). EINVAL is returned for TLS versions that do not match the
original setsockopt call. All are unrecoverable.
strparser is used to parse TLS framing. Decryption is done directly
in to userspace buffers if they are large enough to support it, otherwise
sk_cow_data is called (similar to ipsec), and buffers are decrypted in
place and copied. splice_read always decrypts in place, since no
buffers are provided to decrypt in to.
sk_poll is overridden, and only returns POLLIN if a full TLS message is
received. Otherwise we wait for strparser to finish reading a full frame.
Actual decryption is only done during recvmsg or splice_read calls.
Signed-off-by: Dave Watson <davejwatson@fb.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2018-03-23 00:10:35 +07:00
|
|
|
tls_sw_proto_ops = inet_stream_ops;
|
|
|
|
tls_sw_proto_ops.splice_read = tls_sw_splice_read;
|
|
|
|
|
2018-04-30 14:16:16 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
tls_device_init();
|
|
|
|
#endif
|
2017-06-15 01:37:39 +07:00
|
|
|
tcp_register_ulp(&tcp_tls_ulp_ops);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __exit tls_unregister(void)
|
|
|
|
{
|
|
|
|
tcp_unregister_ulp(&tcp_tls_ulp_ops);
|
2018-04-30 14:16:16 +07:00
|
|
|
#ifdef CONFIG_TLS_DEVICE
|
|
|
|
tls_device_cleanup();
|
|
|
|
#endif
|
2017-06-15 01:37:39 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
module_init(tls_register);
|
|
|
|
module_exit(tls_unregister);
|