summaryrefslogtreecommitdiff
path: root/src/libs/mynewt-nimble/porting/npl/riot/include/nimble
diff options
context:
space:
mode:
authorJF <jf@codingfield.com>2020-04-26 10:25:59 +0200
committerJF <jf@codingfield.com>2020-04-26 10:25:59 +0200
commitbdc10744fb338ae197692713a0b48a7ccc36f566 (patch)
treeaf7a8f2f16ddd2e5483758effec15c7683f6c453 /src/libs/mynewt-nimble/porting/npl/riot/include/nimble
parent032fad094c6411ad3ff4321ad61ceed95d7dc4ff (diff)
Add Nimble in libs directory
Diffstat (limited to 'src/libs/mynewt-nimble/porting/npl/riot/include/nimble')
-rw-r--r--src/libs/mynewt-nimble/porting/npl/riot/include/nimble/nimble_npl_os.h296
1 files changed, 296 insertions, 0 deletions
diff --git a/src/libs/mynewt-nimble/porting/npl/riot/include/nimble/nimble_npl_os.h b/src/libs/mynewt-nimble/porting/npl/riot/include/nimble/nimble_npl_os.h
new file mode 100644
index 00000000..854fd06c
--- /dev/null
+++ b/src/libs/mynewt-nimble/porting/npl/riot/include/nimble/nimble_npl_os.h
@@ -0,0 +1,296 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied. See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+#ifndef _NIMBLE_NPL_OS_H_
+#define _NIMBLE_NPL_OS_H_
+
+#include <stdint.h>
+#include <stdbool.h>
+#include "event/callback.h"
+#include "mutex.h"
+#include "semaphore.h"
+#include "xtimer.h"
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#define BLE_NPL_OS_ALIGNMENT 4
+
+#define BLE_NPL_TIME_FOREVER UINT32_MAX
+
+typedef uint32_t ble_npl_time_t;
+typedef int32_t ble_npl_stime_t;
+
+extern volatile int ble_npl_in_critical;
+
+struct ble_npl_event {
+ event_callback_t e;
+ void *arg;
+};
+
+struct ble_npl_eventq {
+ event_queue_t q;
+};
+
+struct ble_npl_callout {
+ xtimer_t timer;
+ uint64_t target_us;
+ struct ble_npl_event e;
+ event_queue_t *q;
+};
+
+struct ble_npl_mutex {
+ mutex_t mu;
+};
+
+struct ble_npl_sem {
+ sem_t sem;
+};
+
+static inline bool
+ble_npl_os_started(void)
+{
+ return true;
+}
+
+static inline void *
+ble_npl_get_current_task_id(void)
+{
+ return (void *)(uint32_t)thread_getpid();
+}
+
+static inline void
+ble_npl_eventq_init(struct ble_npl_eventq *evq)
+{
+ event_queue_init_detached(&evq->q);
+}
+
+static inline void
+ble_npl_eventq_put(struct ble_npl_eventq *evq, struct ble_npl_event *ev)
+{
+ event_post(&evq->q, &ev->e.super);
+}
+
+static inline struct ble_npl_event *
+ble_npl_eventq_get(struct ble_npl_eventq *evq, ble_npl_time_t tmo)
+{
+ if (evq->q.waiter == NULL) {
+ event_queue_claim(&evq->q);
+ }
+
+ if (tmo == 0) {
+ return (struct ble_npl_event *)event_get(&evq->q);
+ } else if (tmo == BLE_NPL_TIME_FOREVER) {
+ return (struct ble_npl_event *)event_wait(&evq->q);
+ } else {
+ return (struct ble_npl_event *)event_wait_timeout64(&evq->q,
+ tmo * US_PER_MS);
+ }
+}
+
+static inline void
+ble_npl_eventq_remove(struct ble_npl_eventq *evq, struct ble_npl_event *ev)
+{
+ event_cancel(&evq->q, &ev->e.super);
+}
+
+static inline void
+ble_npl_event_run(struct ble_npl_event *ev)
+{
+ ev->e.super.handler(&ev->e.super);
+}
+
+static inline void
+ble_npl_event_init(struct ble_npl_event *ev, ble_npl_event_fn *fn, void *arg)
+{
+ /*
+ * Need to clear list_node manually since init function below does not do
+ * this.
+ */
+ ev->e.super.list_node.next = NULL;
+ event_callback_init(&ev->e, (void(*)(void *))fn, ev);
+ ev->arg = arg;
+}
+
+static inline bool
+ble_npl_event_is_queued(struct ble_npl_event *ev)
+{
+ return (ev->e.super.list_node.next != NULL);
+}
+
+static inline void *
+ble_npl_event_get_arg(struct ble_npl_event *ev)
+{
+ return ev->arg;
+}
+
+static inline void
+ble_npl_event_set_arg(struct ble_npl_event *ev, void *arg)
+{
+ ev->arg = arg;
+}
+
+static inline ble_npl_error_t
+ble_npl_mutex_init(struct ble_npl_mutex *mu)
+{
+ mutex_init(&mu->mu);
+ return BLE_NPL_OK;
+}
+
+static inline ble_npl_error_t
+ble_npl_mutex_pend(struct ble_npl_mutex *mu, ble_npl_time_t timeout)
+{
+ assert(timeout == BLE_NPL_TIME_FOREVER);
+ (void)timeout;
+
+ mutex_lock(&mu->mu);
+ return BLE_NPL_OK;
+}
+
+static inline ble_npl_error_t
+ble_npl_mutex_release(struct ble_npl_mutex *mu)
+{
+ mutex_unlock(&mu->mu);
+ return BLE_NPL_OK;
+}
+
+static inline ble_npl_error_t
+ble_npl_sem_init(struct ble_npl_sem *sem, uint16_t tokens)
+{
+ int rc;
+
+ rc = sem_init(&sem->sem, 0, tokens);
+
+ return rc == 0 ? BLE_NPL_OK : BLE_NPL_ERROR;
+}
+
+static inline ble_npl_error_t
+ble_npl_sem_release(struct ble_npl_sem *sem)
+{
+ int rc;
+
+ rc = sem_post(&sem->sem);
+
+ return rc == 0 ? BLE_NPL_OK : BLE_NPL_ERROR;
+}
+
+static inline uint16_t
+ble_npl_sem_get_count(struct ble_npl_sem *sem)
+{
+ int val = 0;
+
+ sem_getvalue(&sem->sem, &val);
+
+ return (uint16_t)val;
+}
+
+static inline void
+ble_npl_callout_stop(struct ble_npl_callout *co)
+{
+ xtimer_remove(&co->timer);
+}
+
+static inline bool
+ble_npl_callout_is_active(struct ble_npl_callout *c)
+{
+ return (c->timer.offset || c->timer.long_offset);
+}
+
+static inline ble_npl_time_t
+ble_npl_callout_get_ticks(struct ble_npl_callout *co)
+{
+ return (ble_npl_time_t)(co->target_us / US_PER_MS);
+}
+
+static inline void
+ble_npl_callout_set_arg(struct ble_npl_callout *co, void *arg)
+{
+ co->e.arg = arg;
+}
+
+static inline ble_npl_time_t
+ble_npl_time_get(void)
+{
+ return (ble_npl_time_t)(xtimer_now_usec64() / US_PER_MS);
+}
+
+static inline ble_npl_error_t
+ble_npl_time_ms_to_ticks(uint32_t ms, ble_npl_time_t *out_ticks)
+{
+ *out_ticks = ms;
+ return BLE_NPL_OK;
+}
+
+static inline ble_npl_error_t
+ble_npl_time_ticks_to_ms(ble_npl_time_t ticks, uint32_t *out_ms)
+{
+ *out_ms = ticks;
+ return BLE_NPL_OK;
+}
+
+static inline ble_npl_time_t
+ble_npl_time_ms_to_ticks32(uint32_t ms)
+{
+ return ms;
+}
+
+static inline uint32_t
+ble_npl_time_ticks_to_ms32(ble_npl_time_t ticks)
+{
+ return ticks;
+}
+
+static inline void
+ble_npl_time_delay(ble_npl_time_t ticks)
+{
+ xtimer_usleep64(ticks * US_PER_MS);
+}
+
+static inline uint32_t
+ble_npl_hw_enter_critical(void)
+{
+ uint32_t ctx = irq_disable();
+ ++ble_npl_in_critical;
+ return ctx;
+}
+
+static inline void
+ble_npl_hw_exit_critical(uint32_t ctx)
+{
+ --ble_npl_in_critical;
+ irq_restore((unsigned)ctx);
+}
+
+static inline bool
+ble_npl_hw_is_in_critical(void)
+{
+ /*
+ * XXX Currently RIOT does not support an API for finding out if interrupts
+ * are currently disabled, hence in a critical section in this context.
+ * So for now, we use this global variable to keep this state for us.
+ -*/
+ return (ble_npl_in_critical > 0);
+}
+
+#ifdef __cplusplus
+}
+#endif
+
+#endif /* _NPL_H_ */