TOMI: 16- and 32-bit intreadwrite functions
authorMåns Rullgård <mans@mansr.com>
Fri, 19 Feb 2010 03:20:01 +0000 (03:20 +0000)
committerMåns Rullgård <mans@mansr.com>
Fri, 19 Feb 2010 03:20:01 +0000 (03:20 +0000)
Originally committed as revision 21890 to svn://svn.ffmpeg.org/ffmpeg/trunk

libavutil/intreadwrite.h
libavutil/tomi/intreadwrite.h [new file with mode: 0644]

index b1e4e9a..9d029a8 100644 (file)
@@ -60,6 +60,8 @@ typedef union {
 #   include "mips/intreadwrite.h"
 #elif ARCH_PPC
 #   include "ppc/intreadwrite.h"
+#elif ARCH_TOMI
+#   include "tomi/intreadwrite.h"
 #elif ARCH_X86
 #   include "x86/intreadwrite.h"
 #endif
diff --git a/libavutil/tomi/intreadwrite.h b/libavutil/tomi/intreadwrite.h
new file mode 100644 (file)
index 0000000..778b804
--- /dev/null
@@ -0,0 +1,148 @@
+/*
+ * Copyright (c) 2010 Mans Rullgard <mans@mansr.com>
+ *
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef AVUTIL_TOMI_INTREADWRITE_H
+#define AVUTIL_TOMI_INTREADWRITE_H
+
+#include <stdint.h>
+#include "config.h"
+
+#define AV_RB16 AV_RB16
+static av_always_inline uint16_t AV_RB16(const void *p)
+{
+    uint16_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WB16 AV_WB16
+static av_always_inline void AV_WB16(void *p, uint16_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RL16 AV_RL16
+static av_always_inline uint16_t AV_RL16(const void *p)
+{
+    uint16_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WL16 AV_WL16
+static av_always_inline void AV_WL16(void *p, uint16_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RB32 AV_RB32
+static av_always_inline uint32_t AV_RB32(const void *p)
+{
+    uint32_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WB32 AV_WB32
+static av_always_inline void AV_WB32(void *p, uint32_t v)
+{
+    __asm__ volatile ("loadacc,   #4    \n\t"
+                      "add,       %0    \n\t"
+                      "loadacc,   %1    \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RL32 AV_RL32
+static av_always_inline uint32_t AV_RL32(const void *p)
+{
+    uint32_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WL32 AV_WL32
+static av_always_inline void AV_WL32(void *p, uint32_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#endif /* AVUTIL_TOMI_INTREADWRITE_H */