| /* Copyright (c) 2006, Google Inc. |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are |
| * met: |
| * |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following disclaimer |
| * in the documentation and/or other materials provided with the |
| * distribution. |
| * * Neither the name of Google Inc. nor the names of its |
| * contributors may be used to endorse or promote products derived from |
| * this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| * |
| * --- |
| * Author: Sanjay Ghemawat |
| */ |
| |
| #include <stdio.h> |
| #include "base/logging.h" |
| #include "base/atomicops.h" |
| |
| #define GG_ULONGLONG(x) static_cast<uint64>(x) |
| |
| template <class AtomicType> |
| static void TestAtomicIncrement() { |
| // For now, we just test single threaded execution |
| |
| // use a guard value to make sure the NoBarrier_AtomicIncrement doesn't go |
| // outside the expected address bounds. This is in particular to |
| // test that some future change to the asm code doesn't cause the |
| // 32-bit NoBarrier_AtomicIncrement doesn't do the wrong thing on 64-bit |
| // machines. |
| struct { |
| AtomicType prev_word; |
| AtomicType count; |
| AtomicType next_word; |
| } s; |
| |
| AtomicType prev_word_value, next_word_value; |
| memset(&prev_word_value, 0xFF, sizeof(AtomicType)); |
| memset(&next_word_value, 0xEE, sizeof(AtomicType)); |
| |
| s.prev_word = prev_word_value; |
| s.count = 0; |
| s.next_word = next_word_value; |
| |
| ASSERT_EQ(1, base::subtle::NoBarrier_AtomicIncrement(&s.count, 1)); |
| ASSERT_EQ(1, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(3, base::subtle::NoBarrier_AtomicIncrement(&s.count, 2)); |
| ASSERT_EQ(3, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(6, base::subtle::NoBarrier_AtomicIncrement(&s.count, 3)); |
| ASSERT_EQ(6, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(3, base::subtle::NoBarrier_AtomicIncrement(&s.count, -3)); |
| ASSERT_EQ(3, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(1, base::subtle::NoBarrier_AtomicIncrement(&s.count, -2)); |
| ASSERT_EQ(1, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(0, base::subtle::NoBarrier_AtomicIncrement(&s.count, -1)); |
| ASSERT_EQ(0, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(-1, base::subtle::NoBarrier_AtomicIncrement(&s.count, -1)); |
| ASSERT_EQ(-1, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(-5, base::subtle::NoBarrier_AtomicIncrement(&s.count, -4)); |
| ASSERT_EQ(-5, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| |
| ASSERT_EQ(0, base::subtle::NoBarrier_AtomicIncrement(&s.count, 5)); |
| ASSERT_EQ(0, s.count); |
| ASSERT_EQ(prev_word_value, s.prev_word); |
| ASSERT_EQ(next_word_value, s.next_word); |
| } |
| |
| |
| #define NUM_BITS(T) (sizeof(T) * 8) |
| |
| |
| template <class AtomicType> |
| static void TestCompareAndSwap() { |
| AtomicType value = 0; |
| AtomicType prev = base::subtle::NoBarrier_CompareAndSwap(&value, 0, 1); |
| ASSERT_EQ(1, value); |
| ASSERT_EQ(0, prev); |
| |
| // Use test value that has non-zero bits in both halves, more for testing |
| // 64-bit implementation on 32-bit platforms. |
| const AtomicType k_test_val = (GG_ULONGLONG(1) << |
| (NUM_BITS(AtomicType) - 2)) + 11; |
| value = k_test_val; |
| prev = base::subtle::NoBarrier_CompareAndSwap(&value, 0, 5); |
| ASSERT_EQ(k_test_val, value); |
| ASSERT_EQ(k_test_val, prev); |
| |
| value = k_test_val; |
| prev = base::subtle::NoBarrier_CompareAndSwap(&value, k_test_val, 5); |
| ASSERT_EQ(5, value); |
| ASSERT_EQ(k_test_val, prev); |
| } |
| |
| |
| template <class AtomicType> |
| static void TestAtomicExchange() { |
| AtomicType value = 0; |
| AtomicType new_value = base::subtle::NoBarrier_AtomicExchange(&value, 1); |
| ASSERT_EQ(1, value); |
| ASSERT_EQ(0, new_value); |
| |
| // Use test value that has non-zero bits in both halves, more for testing |
| // 64-bit implementation on 32-bit platforms. |
| const AtomicType k_test_val = (GG_ULONGLONG(1) << |
| (NUM_BITS(AtomicType) - 2)) + 11; |
| value = k_test_val; |
| new_value = base::subtle::NoBarrier_AtomicExchange(&value, k_test_val); |
| ASSERT_EQ(k_test_val, value); |
| ASSERT_EQ(k_test_val, new_value); |
| |
| value = k_test_val; |
| new_value = base::subtle::NoBarrier_AtomicExchange(&value, 5); |
| ASSERT_EQ(5, value); |
| ASSERT_EQ(k_test_val, new_value); |
| } |
| |
| |
| template <class AtomicType> |
| static void TestAtomicIncrementBounds() { |
| // Test increment at the half-width boundary of the atomic type. |
| // It is primarily for testing at the 32-bit boundary for 64-bit atomic type. |
| AtomicType test_val = GG_ULONGLONG(1) << (NUM_BITS(AtomicType) / 2); |
| AtomicType value = test_val - 1; |
| AtomicType new_value = base::subtle::NoBarrier_AtomicIncrement(&value, 1); |
| ASSERT_EQ(test_val, value); |
| ASSERT_EQ(value, new_value); |
| |
| base::subtle::NoBarrier_AtomicIncrement(&value, -1); |
| ASSERT_EQ(test_val - 1, value); |
| } |
| |
| // This is a simple sanity check that values are correct. Not testing |
| // atomicity |
| template <class AtomicType> |
| static void TestStore() { |
| const AtomicType kVal1 = static_cast<AtomicType>(0xa5a5a5a5a5a5a5a5LL); |
| const AtomicType kVal2 = static_cast<AtomicType>(-1); |
| |
| AtomicType value; |
| |
| base::subtle::NoBarrier_Store(&value, kVal1); |
| ASSERT_EQ(kVal1, value); |
| base::subtle::NoBarrier_Store(&value, kVal2); |
| ASSERT_EQ(kVal2, value); |
| |
| base::subtle::Acquire_Store(&value, kVal1); |
| ASSERT_EQ(kVal1, value); |
| base::subtle::Acquire_Store(&value, kVal2); |
| ASSERT_EQ(kVal2, value); |
| |
| base::subtle::Release_Store(&value, kVal1); |
| ASSERT_EQ(kVal1, value); |
| base::subtle::Release_Store(&value, kVal2); |
| ASSERT_EQ(kVal2, value); |
| } |
| |
| // This is a simple sanity check that values are correct. Not testing |
| // atomicity |
| template <class AtomicType> |
| static void TestLoad() { |
| const AtomicType kVal1 = static_cast<AtomicType>(0xa5a5a5a5a5a5a5a5LL); |
| const AtomicType kVal2 = static_cast<AtomicType>(-1); |
| |
| AtomicType value; |
| |
| value = kVal1; |
| ASSERT_EQ(kVal1, base::subtle::NoBarrier_Load(&value)); |
| value = kVal2; |
| ASSERT_EQ(kVal2, base::subtle::NoBarrier_Load(&value)); |
| |
| value = kVal1; |
| ASSERT_EQ(kVal1, base::subtle::Acquire_Load(&value)); |
| value = kVal2; |
| ASSERT_EQ(kVal2, base::subtle::Acquire_Load(&value)); |
| |
| value = kVal1; |
| ASSERT_EQ(kVal1, base::subtle::Release_Load(&value)); |
| value = kVal2; |
| ASSERT_EQ(kVal2, base::subtle::Release_Load(&value)); |
| } |
| |
| template <class AtomicType> |
| static void TestAtomicOps() { |
| TestCompareAndSwap<AtomicType>(); |
| TestAtomicExchange<AtomicType>(); |
| TestAtomicIncrementBounds<AtomicType>(); |
| TestStore<AtomicType>(); |
| TestLoad<AtomicType>(); |
| } |
| |
| int main(int argc, char** argv) { |
| TestAtomicIncrement<AtomicWord>(); |
| TestAtomicIncrement<Atomic32>(); |
| |
| TestAtomicOps<AtomicWord>(); |
| TestAtomicOps<Atomic32>(); |
| |
| // I've commented the Atomic64 tests out for now, because Atomic64 |
| // doesn't work on x86 systems that are not compiled to support mmx |
| // registers. Since I want this project to be as portable as |
| // possible -- that is, not to assume we've compiled for mmx or even |
| // that the processor supports it -- and we don't actually use |
| // Atomic64 anywhere, I've commented it out of the test for now. |
| // (Luckily, if we ever do use Atomic64 by accident, we'll get told |
| // via a compiler error rather than some obscure runtime failure, so |
| // this course of action is safe.) |
| // If we ever *do* want to enable this, try adding -msse (or -mmmx?) |
| // to the CXXFLAGS in Makefile.am. |
| #if 0 and defined(BASE_HAS_ATOMIC64) |
| TestAtomicIncrement<base::subtle::Atomic64>(); |
| TestAtomicOps<base::subtle::Atomic64>(); |
| #endif |
| |
| printf("PASS\n"); |
| return 0; |
| } |