aboutsummaryrefslogtreecommitdiffstats
path: root/modules/splat/splat-atomic.c
diff options
context:
space:
mode:
authorbehlendo <behlendo@7e1ea52c-4ff2-0310-8f11-9dd32ca42a1c>2008-03-28 18:21:09 +0000
committerbehlendo <behlendo@7e1ea52c-4ff2-0310-8f11-9dd32ca42a1c>2008-03-28 18:21:09 +0000
commit9f4c835a0efd55139f878c8ed4746cd7da815658 (patch)
tree2cba49e5c08a44483c164441d7f0879a819992d2 /modules/splat/splat-atomic.c
parent4a4295b26736a651a16a2d291868028dbd7cf91b (diff)
Correctly functioning 64-bit atomic shim layer. It's not
what I would call effecient but it does have the advantage of being correct which is all I need right now. I added a regression test as well. git-svn-id: https://outreach.scidac.gov/svn/spl/trunk@57 7e1ea52c-4ff2-0310-8f11-9dd32ca42a1c
Diffstat (limited to 'modules/splat/splat-atomic.c')
-rw-r--r--modules/splat/splat-atomic.c190
1 files changed, 190 insertions, 0 deletions
diff --git a/modules/splat/splat-atomic.c b/modules/splat/splat-atomic.c
new file mode 100644
index 000000000..c170cc05e
--- /dev/null
+++ b/modules/splat/splat-atomic.c
@@ -0,0 +1,190 @@
+#include "splat-internal.h"
+
+#define SPLAT_SUBSYSTEM_ATOMIC 0x0b00
+#define SPLAT_ATOMIC_NAME "atomic"
+#define SPLAT_ATOMIC_DESC "Kernel Atomic Tests"
+
+#define SPLAT_ATOMIC_TEST1_ID 0x0b01
+#define SPLAT_ATOMIC_TEST1_NAME "64-bit"
+#define SPLAT_ATOMIC_TEST1_DESC "Validate 64-bit atomic ops"
+
+#define SPLAT_ATOMIC_TEST_MAGIC 0x43435454UL
+#define SPLAT_ATOMIC_INIT_VALUE 10000000UL
+
+typedef enum {
+ SPLAT_ATOMIC_INC_64 = 0,
+ SPLAT_ATOMIC_DEC_64 = 1,
+ SPLAT_ATOMIC_ADD_64 = 2,
+ SPLAT_ATOMIC_SUB_64 = 3,
+ SPLAT_ATOMIC_ADD_64_NV = 4,
+ SPLAT_ATOMIC_SUB_64_NV = 5,
+ SPLAT_ATOMIC_COUNT_64 = 6
+} atomic_op_t;
+
+typedef struct atomic_priv {
+ unsigned long ap_magic;
+ struct file *ap_file;
+ spinlock_t ap_lock;
+ wait_queue_head_t ap_waitq;
+ volatile uint64_t ap_atomic;
+ volatile uint64_t ap_atomic_exited;
+ atomic_op_t ap_op;
+
+} atomic_priv_t;
+
+static void
+splat_atomic_work(void *priv)
+{
+ atomic_priv_t *ap;
+ atomic_op_t op;
+ int i;
+
+ ap = (atomic_priv_t *)priv;
+ ASSERT(ap->ap_magic == SPLAT_ATOMIC_TEST_MAGIC);
+
+ spin_lock(&ap->ap_lock);
+ op = ap->ap_op;
+ wake_up(&ap->ap_waitq);
+ spin_unlock(&ap->ap_lock);
+
+ splat_vprint(ap->ap_file, SPLAT_ATOMIC_TEST1_NAME,
+ "Thread %d successfully started: %lu/%lu\n", op,
+ (long unsigned)ap->ap_atomic,
+ (long unsigned)ap->ap_atomic_exited);
+
+ for (i = 0; i < SPLAT_ATOMIC_INIT_VALUE / 10; i++) {
+
+ /* Periodically sleep to mix up the ordering */
+ if ((i % (SPLAT_ATOMIC_INIT_VALUE / 100)) == 0) {
+ splat_vprint(ap->ap_file, SPLAT_ATOMIC_TEST1_NAME,
+ "Thread %d sleeping: %lu/%lu\n", op,
+ (long unsigned)ap->ap_atomic,
+ (long unsigned)ap->ap_atomic_exited);
+ set_current_state(TASK_INTERRUPTIBLE);
+ schedule_timeout(HZ / 100);
+ }
+
+ switch (op) {
+ case SPLAT_ATOMIC_INC_64:
+ atomic_inc_64(&ap->ap_atomic);
+ break;
+ case SPLAT_ATOMIC_DEC_64:
+ atomic_dec_64(&ap->ap_atomic);
+ break;
+ case SPLAT_ATOMIC_ADD_64:
+ atomic_add_64(&ap->ap_atomic, 3);
+ break;
+ case SPLAT_ATOMIC_SUB_64:
+ atomic_sub_64(&ap->ap_atomic, 3);
+ break;
+ case SPLAT_ATOMIC_ADD_64_NV:
+ atomic_add_64_nv(&ap->ap_atomic, 5);
+ break;
+ case SPLAT_ATOMIC_SUB_64_NV:
+ atomic_sub_64_nv(&ap->ap_atomic, 5);
+ break;
+ default:
+ BUG_ON(1);
+ }
+ }
+
+ atomic_inc_64(&ap->ap_atomic_exited);
+
+ splat_vprint(ap->ap_file, SPLAT_ATOMIC_TEST1_NAME,
+ "Thread %d successfully exited: %lu/%lu\n", op,
+ (long unsigned)ap->ap_atomic,
+ (long unsigned)ap->ap_atomic_exited);
+
+ thread_exit();
+ wake_up(&ap->ap_waitq);
+}
+
+static int
+splat_atomic_test1_cond(atomic_priv_t *ap)
+{
+ return (ap->ap_atomic_exited == SPLAT_ATOMIC_COUNT_64);
+}
+
+static int
+splat_atomic_test1(struct file *file, void *arg)
+{
+ atomic_priv_t ap;
+ DEFINE_WAIT(wait);
+ kthread_t *thr;
+ int i;
+
+ ap.ap_magic = SPLAT_ATOMIC_TEST_MAGIC;
+ ap.ap_file = file;
+ spin_lock_init(&ap.ap_lock);
+ init_waitqueue_head(&ap.ap_waitq);
+ ap.ap_atomic = SPLAT_ATOMIC_INIT_VALUE;
+ ap.ap_atomic_exited = 0;
+
+ for (i = 0; i < SPLAT_ATOMIC_COUNT_64; i++) {
+ spin_lock(&ap.ap_lock);
+ ap.ap_op = i;
+
+ thr = (kthread_t *)thread_create(NULL, 0, splat_atomic_work,
+ &ap, 0, &p0, TS_RUN,
+ minclsyspri);
+ BUG_ON(thr == NULL);
+
+ /* Prepare to wait, the new thread will wake us once it
+ * has made a copy of the unique private passed data */
+ prepare_to_wait(&ap.ap_waitq, &wait, TASK_UNINTERRUPTIBLE);
+ spin_unlock(&ap.ap_lock);
+ schedule();
+ }
+
+ wait_event_interruptible(ap.ap_waitq, splat_atomic_test1_cond(&ap));
+
+ if (ap.ap_atomic != SPLAT_ATOMIC_INIT_VALUE) {
+ splat_vprint(file, SPLAT_ATOMIC_TEST1_NAME,
+ "Final value %lu does not match initial value %lu\n",
+ (long unsigned)ap.ap_atomic, SPLAT_ATOMIC_INIT_VALUE);
+ return -EINVAL;
+ }
+
+ splat_vprint(file, SPLAT_ATOMIC_TEST1_NAME,
+ "Success initial and final values match, %lu == %lu\n",
+ (long unsigned)ap.ap_atomic, SPLAT_ATOMIC_INIT_VALUE);
+
+ return 0;
+}
+
+splat_subsystem_t *
+splat_atomic_init(void)
+{
+ splat_subsystem_t *sub;
+
+ sub = kmalloc(sizeof(*sub), GFP_KERNEL);
+ if (sub == NULL)
+ return NULL;
+
+ memset(sub, 0, sizeof(*sub));
+ strncpy(sub->desc.name, SPLAT_ATOMIC_NAME, SPLAT_NAME_SIZE);
+ strncpy(sub->desc.desc, SPLAT_ATOMIC_DESC, SPLAT_DESC_SIZE);
+ INIT_LIST_HEAD(&sub->subsystem_list);
+ INIT_LIST_HEAD(&sub->test_list);
+ spin_lock_init(&sub->test_lock);
+ sub->desc.id = SPLAT_SUBSYSTEM_ATOMIC;
+
+ SPLAT_TEST_INIT(sub, SPLAT_ATOMIC_TEST1_NAME, SPLAT_ATOMIC_TEST1_DESC,
+ SPLAT_ATOMIC_TEST1_ID, splat_atomic_test1);
+
+ return sub;
+}
+
+void
+splat_atomic_fini(splat_subsystem_t *sub)
+{
+ ASSERT(sub);
+ SPLAT_TEST_FINI(sub, SPLAT_ATOMIC_TEST1_ID);
+
+ kfree(sub);
+}
+
+int
+splat_atomic_id(void) {
+ return SPLAT_SUBSYSTEM_ATOMIC;
+}