|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
//
|
|
|
|
// See port_example.h for documentation for the following types/functions.
|
|
|
|
|
|
|
|
#ifndef STORAGE_LEVELDB_PORT_PORT_ANDROID_H_
|
|
|
|
#define STORAGE_LEVELDB_PORT_PORT_ANDROID_H_
|
|
|
|
|
|
|
|
#include <endian.h>
|
|
|
|
#include <pthread.h>
|
|
|
|
#include <stdint.h>
|
|
|
|
#include <sha1.h>
|
|
|
|
#include <cstdatomic>
|
|
|
|
#include <string>
|
|
|
|
#include <cctype>
|
|
|
|
|
|
|
|
extern "C" {
|
|
|
|
size_t fread_unlocked(void *a, size_t b, size_t c, FILE *d);
|
|
|
|
size_t fwrite_unlocked(const void *a, size_t b, size_t c, FILE *d);
|
|
|
|
int fflush_unlocked(FILE *f);
|
|
|
|
int fdatasync (int fd);
|
|
|
|
}
|
|
|
|
|
|
|
|
namespace leveldb {
|
|
|
|
namespace port {
|
|
|
|
|
|
|
|
static const bool kLittleEndian = __BYTE_ORDER == __LITTLE_ENDIAN;
|
|
|
|
|
|
|
|
class CondVar;
|
|
|
|
|
|
|
|
class Mutex {
|
|
|
|
public:
|
|
|
|
Mutex();
|
|
|
|
~Mutex();
|
|
|
|
|
|
|
|
void Lock();
|
|
|
|
void Unlock();
|
|
|
|
void AssertHeld() {
|
|
|
|
//TODO(gabor): How can I implement this?
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
friend class CondVar;
|
|
|
|
pthread_mutex_t mu_;
|
|
|
|
|
|
|
|
// No copying
|
|
|
|
Mutex(const Mutex&);
|
|
|
|
void operator=(const Mutex&);
|
|
|
|
};
|
|
|
|
|
|
|
|
class CondVar {
|
|
|
|
public:
|
|
|
|
explicit CondVar(Mutex* mu);
|
|
|
|
~CondVar();
|
|
|
|
void Wait();
|
|
|
|
void Signal();
|
|
|
|
void SignalAll();
|
|
|
|
private:
|
|
|
|
Mutex* mu_;
|
|
|
|
pthread_cond_t cv_;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Storage for a lock-free pointer
|
|
|
|
class AtomicPointer {
|
|
|
|
private:
|
|
|
|
std::atomic<void*> rep_;
|
|
|
|
public:
|
|
|
|
AtomicPointer() { }
|
|
|
|
explicit AtomicPointer(void* v) : rep_(v) { }
|
|
|
|
inline void* Acquire_Load() const {
|
|
|
|
return rep_.load(std::memory_order_acquire);
|
|
|
|
}
|
|
|
|
inline void Release_Store(void* v) {
|
|
|
|
rep_.store(v, std::memory_order_release);
|
|
|
|
}
|
|
|
|
inline void* NoBarrier_Load() const {
|
|
|
|
return rep_.load(std::memory_order_relaxed);
|
|
|
|
}
|
|
|
|
inline void NoBarrier_Store(void* v) {
|
|
|
|
rep_.store(v, std::memory_order_relaxed);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// TODO(gabor): Implement actual compress
|
|
|
|
inline bool Snappy_Compress(
|
|
|
|
const char* input,
|
|
|
|
size_t input_length,
|
|
|
|
std::string* output) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(gabor): Implement actual uncompress
|
|
|
|
inline bool Snappy_Uncompress(
|
|
|
|
const char* input_data,
|
|
|
|
size_t input_length,
|
|
|
|
std::string* output) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
inline void SHA1_Hash(const char* data, size_t len, char* hash_array) {
|
|
|
|
SHA1_CTX sha1_ctx;
|
|
|
|
SHA1Init(&sha1_ctx);
|
|
|
|
SHA1Update(&sha1_ctx, (const u_char*)data, len);
|
|
|
|
SHA1Final((u_char*)hash_array, &sha1_ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
inline uint64_t ThreadIdentifier() {
|
|
|
|
pthread_t tid = pthread_self();
|
|
|
|
uint64_t r = 0;
|
|
|
|
memcpy(&r, &tid, sizeof(r) < sizeof(tid) ? sizeof(r) : sizeof(tid));
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
inline bool GetHeapProfile(void (*func)(void*, const char*, int), void* arg) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // STORAGE_LEVELDB_PORT_PORT_ANDROID_H_
|