Re: [PATCH v6 3/4] tee: optee: support tracking system threads

From: Etienne Carriere
Date: Thu May 11 2023 - 03:25:14 EST


Hi Sumit,


On Wed, 10 May 2023 at 19:38, Etienne Carriere
<etienne.carriere@xxxxxxxxxx> wrote:
>
> On Wed, 10 May 2023 at 17:15, Etienne Carriere
> <etienne.carriere@xxxxxxxxxx> wrote:
> >
> > On Wed, 10 May 2023 at 12:08, Sumit Garg <sumit.garg@xxxxxxxxxx> wrote:
> > >
> > > On Fri, 5 May 2023 at 23:01, Etienne Carriere
> > > <etienne.carriere@xxxxxxxxxx> wrote:
> > > >
> > > > From: Jens Wiklander <jens.wiklander@xxxxxxxxxx>
> > > >
> > > > Adds support in the OP-TEE driver to keep track of reserved system
> > > > threads. The optee_cq_*() functions are updated to handle this if
> > > > enabled. The SMC ABI part of the driver enables this tracking, but the
> > > > FF-A ABI part does not.
> > >
> > > OP-TEE system threads sound like a core feature towards OP-TEE. If we
> > > enable it only for SMC ABI then it is likely to break kernel drivers
> > > who migrate to FFA ABI. Also, looking from implementation point of
> > > view it shouldn't be that hard to enable it for FFA ABI too.
> >
> > It is not mandatory all TEE ABIs support this feature.
> > Caller driver can still use SMC/OP-TEE transport without this support
> > yet with some limitation of course.
> >
> > >
> > > >
> > > > Signed-off-by: Jens Wiklander <jens.wiklander@xxxxxxxxxx>
> > > > Signed-off-by: Etienne Carriere <etienne.carriere@xxxxxxxxxx>
> > > > ---
> > > > No change since v5
> > > >
> > > > Changes since v4:
> > > > - New change that supersedes implementation proposed in PATCH v4
> > > > (tee: system invocation"). Thanks to Jens implementation we don't need
> > > > the new OP-TEE services that my previous patch versions introduced to
> > > > monitor system threads entry. Now, Linux optee SMC ABI driver gets TEE
> > > > provisioned thread contexts count once and monitors thread entries in
> > > > OP-TEE on that basis and the system thread capability of the related
> > > > tee session. By the way, I dropped the WARN_ONCE() call I suggested
> > > > on tee thread exhaustion as it does not provides useful information.
> > > > ---
> > > > drivers/tee/optee/call.c | 128 +++++++++++++++++++++++++++---
> > > > drivers/tee/optee/ffa_abi.c | 10 +--
> > > > drivers/tee/optee/optee_private.h | 13 ++-
> > > > drivers/tee/optee/smc_abi.c | 24 ++++--
> > > > 4 files changed, 154 insertions(+), 21 deletions(-)
> > > >
> > > > diff --git a/drivers/tee/optee/call.c b/drivers/tee/optee/call.c
> > > > index dba5339b61ae..c2d484201f79 100644
> > > > --- a/drivers/tee/optee/call.c
> > > > +++ b/drivers/tee/optee/call.c
> > > > @@ -39,9 +39,26 @@ struct optee_shm_arg_entry {
> > > > DECLARE_BITMAP(map, MAX_ARG_COUNT_PER_ENTRY);
> > > > };
> > > >
> > > > +void optee_cq_init(struct optee_call_queue *cq, int thread_count)
> > > > +{
> > > > + mutex_init(&cq->mutex);
> > > > + INIT_LIST_HEAD(&cq->normal_waiters);
> > > > + INIT_LIST_HEAD(&cq->sys_waiters);
> > > > + /*
> > > > + * If cq->total_thread_count is 0 then we're not trying to keep
> > > > + * track of how many free threads we have, instead we're relying on
> > > > + * the secure world to tell us when we're out of thread and have to
> > > > + * wait for another thread to become available.
> > > > + */
> > > > + cq->total_thread_count = thread_count;
> > > > + cq->free_normal_thread_count = thread_count;
> > > > +}
> > > > +
> > > > void optee_cq_wait_init(struct optee_call_queue *cq,
> > > > - struct optee_call_waiter *w)
> > > > + struct optee_call_waiter *w, bool sys_thread)
> > >
> > > Introduction of system_thread property should be part of patch #1.
> >
> > Patch 1 prepared for TEE system thread support.
> > This patch implements it for the SMC ABI.
> > This split looked easier from a patch review perspective.
>
> After a second look, I understand your comment.
> Right, i'll move this function ABI change to patch #1 and ditto for
> the other places your pointed out.

Actually, after making changes, it does not look better.
Nevertheless, I'll post a patch v7 with your comment address. You will
see if it makes changes more clear.

br,
etienne

>
> >
> >
> > >
> > > > {
> > > > + bool need_wait = false;
> > > > +
> > > > /*
> > > > * We're preparing to make a call to secure world. In case we can't
> > > > * allocate a thread in secure world we'll end up waiting in
> > > > @@ -53,15 +70,40 @@ void optee_cq_wait_init(struct optee_call_queue *cq,
> > > > mutex_lock(&cq->mutex);
> > > >
> > > > /*
> > > > - * We add ourselves to the queue, but we don't wait. This
> > > > - * guarantees that we don't lose a completion if secure world
> > > > - * returns busy and another thread just exited and try to complete
> > > > - * someone.
> > > > + * We add ourselves to a queue, but we don't wait. This guarantees
> > > > + * that we don't lose a completion if secure world returns busy and
> > > > + * another thread just exited and try to complete someone.
> > > > */
> > > > init_completion(&w->c);
> > > > - list_add_tail(&w->list_node, &cq->waiters);
> > > > + w->sys_thread = sys_thread;
> > > > + if (sys_thread) {
> > > > + list_add_tail(&w->list_node, &cq->sys_waiters);
> > > > + } else {
> > > > + list_add_tail(&w->list_node, &cq->normal_waiters);
> > > > + if (cq->total_thread_count) {
> > > > + /*
> > > > + * Claim a normal thread if one is available, else
> > > > + * we'll need to wait for a normal thread to be
> > > > + * released.
> > > > + */
> > > > + if (cq->free_normal_thread_count > 0)
> > > > + cq->free_normal_thread_count--;
> > > > + else
> > > > + need_wait = true;
> > > > + }
> > > > + }
> > > >
> > > > mutex_unlock(&cq->mutex);
> > > > +
> > > > + while (need_wait) {
> > > > + optee_cq_wait_for_completion(cq, w);
> > > > + mutex_lock(&cq->mutex);
> > > > + if (cq->free_normal_thread_count > 0) {
> > > > + cq->free_normal_thread_count--;
> > > > + need_wait = false;
> > > > + }
> > > > + mutex_unlock(&cq->mutex);
> > > > + }
> > > > }
> > > >
> > > > void optee_cq_wait_for_completion(struct optee_call_queue *cq,
> > > > @@ -74,7 +116,10 @@ void optee_cq_wait_for_completion(struct optee_call_queue *cq,
> > > > /* Move to end of list to get out of the way for other waiters */
> > > > list_del(&w->list_node);
> > > > reinit_completion(&w->c);
> > > > - list_add_tail(&w->list_node, &cq->waiters);
> > > > + if (w->sys_thread)
> > > > + list_add_tail(&w->list_node, &cq->sys_waiters);
> > > > + else
> > > > + list_add_tail(&w->list_node, &cq->normal_waiters);
> > > >
> > > > mutex_unlock(&cq->mutex);
> > > > }
> > > > @@ -83,10 +128,19 @@ static void optee_cq_complete_one(struct optee_call_queue *cq)
> > > > {
> > > > struct optee_call_waiter *w;
> > > >
> > > > - list_for_each_entry(w, &cq->waiters, list_node) {
> > > > + list_for_each_entry(w, &cq->sys_waiters, list_node) {
> > > > if (!completion_done(&w->c)) {
> > > > complete(&w->c);
> > > > - break;
> > > > + return;
> > > > + }
> > > > + }
> > > > +
> > > > + if (!cq->total_thread_count || cq->free_normal_thread_count > 0) {
> > > > + list_for_each_entry(w, &cq->normal_waiters, list_node) {
> > > > + if (!completion_done(&w->c)) {
> > > > + complete(&w->c);
> > > > + break;
> > > > + }
> > > > }
> > > > }
> > > > }
> > > > @@ -104,6 +158,9 @@ void optee_cq_wait_final(struct optee_call_queue *cq,
> > > > /* Get out of the list */
> > > > list_del(&w->list_node);
> > > >
> > > > + if (!w->sys_thread)
> > > > + cq->free_normal_thread_count++; /* Release a normal thread */
> > > > +
> > > > /* Wake up one eventual waiting task */
> > > > optee_cq_complete_one(cq);
> > > >
> > > > @@ -119,6 +176,36 @@ void optee_cq_wait_final(struct optee_call_queue *cq,
> > > > mutex_unlock(&cq->mutex);
> > > > }
> > > >
> > > > +bool optee_cq_inc_sys_thread_count(struct optee_call_queue *cq)
> > > > +{
> > > > + bool rc = false;
> > > > +
> > > > + mutex_lock(&cq->mutex);
> > > > +
> > > > + /* Leave at least 1 normal (non-system) thread */
> > >
> > > IMO, this might be counter productive. As most kernel drivers open a
> > > session during driver probe which are only released in the driver
> > > release method.
> >
> > It is always the case?
>
> This answer of mine is irrelevant. Sorry,
> Please read only the below comments of mine, especially:
> | Note that an OP-TEE thread is not bound to a TEE session but rather
> | bound to a yielded call to OP-TEE.
>
> >
> > > If the kernel driver is built-in then the session is
> > > never released. Now with system threads we would reserve an OP-TEE
> > > thread for that kernel driver as well which will never be available to
> > > regular user-space clients.
> >
> > That is not true. No driver currently requests their TEE thread to be
> > a system thread.
> > Only SCMI does because it needs to by construction.
> >
> >
> > > So I would rather suggest we only allow a
> > > single system thread to be reserved as a starting point which is
> > > relevant to this critical SCMI service. We can also make this upper
> > > bound for system threads configurable with default value as 1 if
> > > needed.
>
> Note that SCMI server can expose several SCMI channels (at most 1 per
> SCMI protocol used) and each of them will need to request a
> system_thread to TEE driver.
>
> Etienne
>
> >
> > Reserving one or more system threads depends on the number of thread
> > context provisioned by the TEE.
> > Note that the implementation proposed here prevents Linux kernel from
> > exhausting TEE threads so user space always has at least a TEE thread
> > context left available.
> >
> > Note that an OP-TEE thread is not bound to a TEE session but rather
> > bound to a yielded call to OP-TEE.
> >
> >
> > >
> > > > + if (cq->res_sys_thread_count + 1 < cq->total_thread_count) {
> > > > + cq->free_normal_thread_count--;
> > > > + cq->res_sys_thread_count++;
> > > > + rc = true;
> > > > + }
> > > > +
> > > > + mutex_unlock(&cq->mutex);
> > > > +
> > > > + return rc;
> > > > +}
> > > > +
> > > > +void optee_cq_dec_sys_thread_count(struct optee_call_queue *cq)
> > > > +{
> > > > + mutex_lock(&cq->mutex);
> > > > + if (cq->res_sys_thread_count > 0) {
> > > > + cq->res_sys_thread_count--;
> > > > + cq->free_normal_thread_count++;
> > > > + /* If there's someone waiting, let it resume */
> > > > + optee_cq_complete_one(cq);
> > > > + }
> > > > + mutex_unlock(&cq->mutex);
> > > > +}
> > > > +
> > > > /* Requires the filpstate mutex to be held */
> > > > static struct optee_session *find_session(struct optee_context_data *ctxdata,
> > > > u32 session_id)
> > > > @@ -361,6 +448,27 @@ int optee_open_session(struct tee_context *ctx,
> > > > return rc;
> > > > }
> > > >
> > > > +int optee_system_session(struct tee_context *ctx, u32 session)
> > > > +{
> > > > + struct optee_context_data *ctxdata = ctx->data;
> > > > + struct optee *optee = tee_get_drvdata(ctx->teedev);
> > > > + struct optee_session *sess;
> > > > + int rc = -EINVAL;
> > > > +
> > > > + mutex_lock(&ctxdata->mutex);
> > > > +
> > > > + sess = find_session(ctxdata, session);
> > > > + if (sess && !sess->use_sys_thread &&
> > > > + optee_cq_inc_sys_thread_count(&optee->call_queue)) {
> > > > + rc = 0;
> > > > + sess->use_sys_thread = true;
> > > > + }
> > > > +
> > > > + mutex_unlock(&ctxdata->mutex);
> > > > +
> > > > + return rc;
> > > > +}
> > > > +
> > > > int optee_close_session_helper(struct tee_context *ctx, u32 session,
> > > > bool system_thread)
> > > > {
> > > > @@ -378,6 +486,8 @@ int optee_close_session_helper(struct tee_context *ctx, u32 session,
> > > > msg_arg->session = session;
> > > > optee->ops->do_call_with_arg(ctx, shm, offs, system_thread);
> > > >
> > > > + if (system_thread)
> > > > + optee_cq_dec_sys_thread_count(&optee->call_queue);
> > > > optee_free_msg_arg(ctx, entry, offs);
> > > >
> > > > return 0;
> > > > diff --git a/drivers/tee/optee/ffa_abi.c b/drivers/tee/optee/ffa_abi.c
> > > > index 52cec9d06041..0c9055691343 100644
> > > > --- a/drivers/tee/optee/ffa_abi.c
> > > > +++ b/drivers/tee/optee/ffa_abi.c
> > > > @@ -528,7 +528,8 @@ static void optee_handle_ffa_rpc(struct tee_context *ctx, struct optee *optee,
> > > >
> > > > static int optee_ffa_yielding_call(struct tee_context *ctx,
> > > > struct ffa_send_direct_data *data,
> > > > - struct optee_msg_arg *rpc_arg)
> > > > + struct optee_msg_arg *rpc_arg,
> > > > + bool system_thread)
> > > > {
> > > > struct optee *optee = tee_get_drvdata(ctx->teedev);
> > > > struct ffa_device *ffa_dev = optee->ffa.ffa_dev;
> > > > @@ -541,7 +542,7 @@ static int optee_ffa_yielding_call(struct tee_context *ctx,
> > > > int rc;
> > > >
> > > > /* Initialize waiter */
> > > > - optee_cq_wait_init(&optee->call_queue, &w);
> > > > + optee_cq_wait_init(&optee->call_queue, &w, system_thread);
> > > > while (true) {
> > > > rc = msg_ops->sync_send_receive(ffa_dev, data);
> > > > if (rc)
> > > > @@ -643,7 +644,7 @@ static int optee_ffa_do_call_with_arg(struct tee_context *ctx,
> > > > if (IS_ERR(rpc_arg))
> > > > return PTR_ERR(rpc_arg);
> > > >
> > > > - return optee_ffa_yielding_call(ctx, &data, rpc_arg);
> > > > + return optee_ffa_yielding_call(ctx, &data, rpc_arg, system_thread);
> > > > }
> > >
> > > Introduction of system_thread property should be part of patch #1.
> >
> > See my answer above.
> >
> > >
> > > >
> > > > /*
> > > > @@ -851,8 +852,7 @@ static int optee_ffa_probe(struct ffa_device *ffa_dev)
> > > > if (rc)
> > > > goto err_unreg_supp_teedev;
> > > > mutex_init(&optee->ffa.mutex);
> > > > - mutex_init(&optee->call_queue.mutex);
> > > > - INIT_LIST_HEAD(&optee->call_queue.waiters);
> > > > + optee_cq_init(&optee->call_queue, 0);
> > >
> > > This looks like some refactoring going on which should be part of a
> > > separate patch to ease the review process.
> >
> > I'll see how to handle your request.
> >
> > >
> > > > optee_supp_init(&optee->supp);
> > > > optee_shm_arg_cache_init(optee, arg_cache_flags);
> > > > ffa_dev_set_drvdata(ffa_dev, optee);
> > > > diff --git a/drivers/tee/optee/optee_private.h b/drivers/tee/optee/optee_private.h
> > > > index 3da7960ab34a..6e0863a70843 100644
> > > > --- a/drivers/tee/optee/optee_private.h
> > > > +++ b/drivers/tee/optee/optee_private.h
> > > > @@ -43,12 +43,17 @@ typedef void (optee_invoke_fn)(unsigned long, unsigned long, unsigned long,
> > > > struct optee_call_waiter {
> > > > struct list_head list_node;
> > > > struct completion c;
> > > > + bool sys_thread;
> > > > };
> > > >
> > > > struct optee_call_queue {
> > > > /* Serializes access to this struct */
> > > > struct mutex mutex;
> > > > - struct list_head waiters;
> > > > + struct list_head normal_waiters;
> > > > + struct list_head sys_waiters;
> > > > + int total_thread_count;
> > > > + int free_normal_thread_count;
> > > > + int res_sys_thread_count;
> > > > };
> > > >
> > > > struct optee_notif {
> > > > @@ -254,6 +259,7 @@ int optee_supp_send(struct tee_context *ctx, u32 ret, u32 num_params,
> > > > int optee_open_session(struct tee_context *ctx,
> > > > struct tee_ioctl_open_session_arg *arg,
> > > > struct tee_param *param);
> > > > +int optee_system_session(struct tee_context *ctx, u32 session);
> > > > int optee_close_session_helper(struct tee_context *ctx, u32 session,
> > > > bool system_thread);
> > > > int optee_close_session(struct tee_context *ctx, u32 session);
> > > > @@ -303,8 +309,11 @@ static inline void optee_to_msg_param_value(struct optee_msg_param *mp,
> > > > mp->u.value.c = p->u.value.c;
> > > > }
> > > >
> > > > +void optee_cq_init(struct optee_call_queue *cq, int thread_count);
> > > > +bool optee_cq_inc_sys_thread_count(struct optee_call_queue *cq);
> > > > +void optee_cq_dec_sys_thread_count(struct optee_call_queue *cq);
> > > > void optee_cq_wait_init(struct optee_call_queue *cq,
> > > > - struct optee_call_waiter *w);
> > > > + struct optee_call_waiter *w, bool sys_thread);
> > > > void optee_cq_wait_for_completion(struct optee_call_queue *cq,
> > > > struct optee_call_waiter *w);
> > > > void optee_cq_wait_final(struct optee_call_queue *cq,
> > > > diff --git a/drivers/tee/optee/smc_abi.c b/drivers/tee/optee/smc_abi.c
> > > > index 56ebbb96ac97..2819674fd555 100644
> > > > --- a/drivers/tee/optee/smc_abi.c
> > > > +++ b/drivers/tee/optee/smc_abi.c
> > > > @@ -281,9 +281,10 @@ static int optee_to_msg_param(struct optee *optee,
> > > > static void optee_enable_shm_cache(struct optee *optee)
> > > > {
> > > > struct optee_call_waiter w;
> > > > + bool system_thread = false;
> > >
> > > This variable is redundant.
> >
> > Using a variable here makes it more clear which argument is passed to
> > optee_cq_wait_init().
> > Calling optee_cq_wait_init(&optee->call_queue, &w, false); is less readable.
> >
> > >
> > > >
> > > > /* We need to retry until secure world isn't busy. */
> > > > - optee_cq_wait_init(&optee->call_queue, &w);
> > > > + optee_cq_wait_init(&optee->call_queue, &w, system_thread);
> > > > while (true) {
> > > > struct arm_smccc_res res;
> > > >
> > > > @@ -306,9 +307,10 @@ static void optee_enable_shm_cache(struct optee *optee)
> > > > static void __optee_disable_shm_cache(struct optee *optee, bool is_mapped)
> > > > {
> > > > struct optee_call_waiter w;
> > > > + bool system_thread = false;
> > > >
> > >
> > > This variable is redundant.
> >
> > Ditto
> >
> > >
> > > > /* We need to retry until secure world isn't busy. */
> > > > - optee_cq_wait_init(&optee->call_queue, &w);
> > > > + optee_cq_wait_init(&optee->call_queue, &w, system_thread);
> > > > while (true) {
> > > > union {
> > > > struct arm_smccc_res smccc;
> > > > @@ -927,7 +929,7 @@ static int optee_smc_do_call_with_arg(struct tee_context *ctx,
> > > > reg_pair_from_64(&param.a1, &param.a2, parg);
> > > > }
> > > > /* Initialize waiter */
> > > > - optee_cq_wait_init(&optee->call_queue, &w);
> > > > + optee_cq_wait_init(&optee->call_queue, &w, system_thread);
> > > > while (true) {
> > > > struct arm_smccc_res res;
> > > >
> > > > @@ -1209,6 +1211,7 @@ static const struct tee_driver_ops optee_clnt_ops = {
> > > > .release = optee_release,
> > > > .open_session = optee_open_session,
> > > > .close_session = optee_close_session,
> > > > + .system_session = optee_system_session,
> > > > .invoke_func = optee_invoke_func,
> > > > .cancel_req = optee_cancel_req,
> > > > .shm_register = optee_shm_register,
> > > > @@ -1356,6 +1359,16 @@ static bool optee_msg_exchange_capabilities(optee_invoke_fn *invoke_fn,
> > > > return true;
> > > > }
> > > >
> > > > +static unsigned int optee_msg_get_thread_count(optee_invoke_fn *invoke_fn)
> > > > +{
> > > > + struct arm_smccc_res res;
> > > > +
> > > > + invoke_fn(OPTEE_SMC_GET_THREAD_COUNT, 0, 0, 0, 0, 0, 0, 0, &res);
> > > > + if (res.a0)
> > > > + return 0;
> > > > + return res.a1;
> > > > +}
> > > > +
> > > > static struct tee_shm_pool *
> > > > optee_config_shm_memremap(optee_invoke_fn *invoke_fn, void **memremaped_shm)
> > > > {
> > > > @@ -1609,6 +1622,7 @@ static int optee_probe(struct platform_device *pdev)
> > > > struct optee *optee = NULL;
> > > > void *memremaped_shm = NULL;
> > > > unsigned int rpc_param_count;
> > > > + unsigned int thread_count;
> > > > struct tee_device *teedev;
> > > > struct tee_context *ctx;
> > > > u32 max_notif_value;
> > > > @@ -1636,6 +1650,7 @@ static int optee_probe(struct platform_device *pdev)
> > > > return -EINVAL;
> > > > }
> > > >
> > > > + thread_count = optee_msg_get_thread_count(invoke_fn);
> > > > if (!optee_msg_exchange_capabilities(invoke_fn, &sec_caps,
> > > > &max_notif_value,
> > > > &rpc_param_count)) {
> > > > @@ -1725,8 +1740,7 @@ static int optee_probe(struct platform_device *pdev)
> > > > if (rc)
> > > > goto err_unreg_supp_teedev;
> > > >
> > > > - mutex_init(&optee->call_queue.mutex);
> > > > - INIT_LIST_HEAD(&optee->call_queue.waiters);
> > > > + optee_cq_init(&optee->call_queue, thread_count);
> > >
> > > Again, this looks like some refactoring going on which should be part
> > > of a separate patch to ease the review process.
> >
> > I get your point.
> > Thanks again for the feedback.
> >
> > Br,
> > etienne
> >
> > >
> > > -Sumit
> > >
> > > > optee_supp_init(&optee->supp);
> > > > optee->smc.memremaped_shm = memremaped_shm;
> > > > optee->pool = pool;
> > > > --
> > > > 2.25.1
> > > >