blob: 1328e35f865866842218a831e7a75f1babfcd4ef [file] [log] [blame]
// Copyright 2016 The SwiftShader Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#ifndef sw_MutexLock_hpp
#define sw_MutexLock_hpp
#include "Thread.hpp"
#if defined(__linux__)
// Use a pthread mutex on Linux. Since many processes may use SwiftShader
// at the same time it's best to just have the scheduler overhead.
#include <pthread.h>
namespace sw
{
class MutexLock
{
public:
MutexLock()
{
pthread_mutex_init(&mutex, NULL);
}
~MutexLock()
{
pthread_mutex_destroy(&mutex);
}
bool attemptLock()
{
return pthread_mutex_trylock(&mutex) == 0;
}
void lock()
{
pthread_mutex_lock(&mutex);
}
void unlock()
{
pthread_mutex_unlock(&mutex);
}
private:
pthread_mutex_t mutex;
};
}
#else // !__linux__
#include <atomic>
namespace sw
{
class BackoffLock
{
public:
BackoffLock()
{
mutex = 0;
}
bool attemptLock()
{
if(!isLocked())
{
if(mutex.exchange(true) == false)
{
return true;
}
}
return false;
}
void lock()
{
int backoff = 1;
while(!attemptLock())
{
if(backoff <= 64)
{
for(int i = 0; i < backoff; i++)
{
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
nop();
}
backoff *= 2;
}
else
{
Thread::yield();
backoff = 1;
}
};
}
void unlock()
{
mutex.store(false, std::memory_order_release);
}
bool isLocked()
{
return mutex.load(std::memory_order_acquire);
}
private:
struct
{
// Ensure that the mutex variable is on its own 64-byte cache line to avoid false sharing
// Padding must be public to avoid compiler warnings
volatile int padding1[16];
std::atomic<bool> mutex;
volatile int padding2[15];
};
};
using MutexLock = BackoffLock;
}
#endif // !__linux__
class LockGuard
{
public:
explicit LockGuard(sw::MutexLock &mutex) : mutex(&mutex)
{
mutex.lock();
}
explicit LockGuard(sw::MutexLock *mutex) : mutex(mutex)
{
if (mutex) mutex->lock();
}
~LockGuard()
{
if (mutex) mutex->unlock();
}
protected:
sw::MutexLock *mutex;
};
#endif // sw_MutexLock_hpp