* invalidate any other reasons why the executable file might be covered by
* the GNU General Public License.
*
- * Copyright 2009 Develer S.r.l. (http://www.develer.com/)
+ * Copyright 2010 Develer S.r.l. (http://www.develer.com/)
*
* -->
*
#include "kblock.h"
+#define LOG_LEVEL LOG_LVL_ERR
+#define LOG_FORMAT LOG_FMT_VERBOSE
-static void *kblock_swMap(struct KBlock *b, size_t offset, size_t size)
+#include <cfg/log.h>
+#include <string.h>
+
+INLINE size_t kblock_readDirect(struct KBlock *b, block_idx_t index, void *buf, size_t offset, size_t size)
{
- return (kblock_readBuf(b, b->priv.pagebuf, offset, size) == size) ? b->priv.pagebuf : NULL;
+ KB_ASSERT_METHOD(b, readDirect);
+ ASSERT(index < b->blk_cnt);
+ return b->priv.vt->readDirect(b, b->priv.blk_start + index, buf, offset, size);
}
+INLINE size_t kblock_writeDirect(struct KBlock *b, block_idx_t index, const void *buf, size_t offset, size_t size)
+{
+ KB_ASSERT_METHOD(b, writeDirect);
+ ASSERT(index < b->blk_cnt);
+ return b->priv.vt->writeDirect(b, b->priv.blk_start + index, buf, offset, size);
+}
-static int kblock_swUnmap(struct KBlock *b, size_t offset, size_t size)
+INLINE size_t kblock_readBuf(struct KBlock *b, void *buf, size_t offset, size_t size)
{
- return (kblock_writeBuf(b, b->priv.pagebuf, offset, size) == size) ? 0 : EOF;
+ KB_ASSERT_METHOD(b, readBuf);
+ ASSERT(offset + size <= b->blk_size);
+
+ return b->priv.vt->readBuf(b, buf, offset, size);
}
+INLINE size_t kblock_writeBuf(struct KBlock *b, const void *buf, size_t offset, size_t size)
+{
+ KB_ASSERT_METHOD(b, writeBuf);
+ ASSERT(offset + size <= b->blk_size);
+ return b->priv.vt->writeBuf(b, buf, offset, size);
+}
-void *kblock_unsupportedMap(struct KBlock *b, UNUSED_ARG(size_t, offset), UNUSED_ARG(size_t, size))
+INLINE int kblock_load(struct KBlock *b, block_idx_t index)
{
- LOG_WARN("This driver does not support block mapping: use kblock_addMapping() to add generic mapping functionality.\n");
- b->priv.flags |= BV(KBS_ERR_MAP_NOT_AVAILABLE);
- return NULL;
+ KB_ASSERT_METHOD(b, load);
+ ASSERT(index < b->blk_cnt);
+
+ LOG_INFO("index %ld\n", index);
+ return b->priv.vt->load(b, b->priv.blk_start + index);
}
-void kblock_addMapping(struct KBlock *dev, void *buf, size_t size)
+INLINE int kblock_store(struct KBlock *b, block_idx_t index)
{
+ KB_ASSERT_METHOD(b, store);
+ ASSERT(index < b->blk_cnt);
+
+ LOG_INFO("index %ld\n", index);
+ return b->priv.vt->store(b, b->priv.blk_start + index);
+}
+
+INLINE void kblock_setDirty(struct KBlock *b, bool dirty)
+{
+ if (dirty)
+ b->priv.flags |= KB_CACHE_DIRTY;
+ else
+ b->priv.flags &= ~KB_CACHE_DIRTY;
+}
+
+
+size_t kblock_read(struct KBlock *b, block_idx_t idx, void *buf, size_t offset, size_t size)
+{
+ ASSERT(b);
ASSERT(buf);
- ASSERT(size);
- ASSERT(dev);
-
- dev->vt->map = kblock_swMap;
- dev->vt->unmap = kblock_swUnmap;
-
- dev->priv.pagebuf = buf;
- dev->priv.pagebuf_size = size;
+ ASSERT(offset + size <= b->blk_size);
+ LOG_INFO("blk_idx %ld, offset %u, size %u\n", idx, offset, size);
+
+ if (kblock_buffered(b) && idx == b->priv.curr_blk)
+ return kblock_readBuf(b, buf, offset, size);
+ else
+ return kblock_readDirect(b, idx, buf, offset, size);
+}
+
+
+int kblock_flush(struct KBlock *b)
+{
+ ASSERT(b);
+
+ if (!kblock_buffered(b))
+ return 0;
+
+ if (kblock_cacheDirty(b))
+ {
+ LOG_INFO("flushing block %ld\n", b->priv.curr_blk);
+ if (kblock_store(b, b->priv.curr_blk) == 0)
+ kblock_setDirty(b, false);
+ else
+ return EOF;
+ }
+ return 0;
+}
+
+
+static bool kblock_loadPage(struct KBlock *b, block_idx_t idx)
+{
+ ASSERT(b);
+
+ if (idx != b->priv.curr_blk)
+ {
+ LOG_INFO("loading block %ld\n", idx);
+ if (kblock_flush(b) != 0 || kblock_load(b, idx) != 0)
+ return false;
+
+ b->priv.curr_blk = idx;
+ }
+ return true;
+}
+
+
+int kblock_trim(struct KBlock *b, block_idx_t start, block_idx_t count)
+{
+ ASSERT(start + count <= b->blk_cnt);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, start))
+ return EOF;
+ }
+
+ b->priv.blk_start += start;
+ b->priv.curr_blk = 0; // adjust logical address
+ b->blk_cnt = count;
+ return 0;
+}
+
+
+size_t kblock_write(struct KBlock *b, block_idx_t idx, const void *buf, size_t offset, size_t size)
+{
+ ASSERT(b);
+ ASSERT(buf);
+ ASSERT(idx < b->blk_cnt);
+ ASSERT(offset + size <= b->blk_size);
+
+ LOG_INFO("blk_idx %ld, offset %u, size %u\n", idx, offset, size);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, idx))
+ return 0;
+
+ kblock_setDirty(b, true);
+ return kblock_writeBuf(b, buf, offset, size);
+ }
+ else
+ {
+ #ifdef _DEBUG
+ if (offset != 0 || size != b->blk_size)
+ ASSERT(kblock_partialWrite(b));
+ #endif
+ return kblock_writeDirect(b, idx, buf, offset, size);
+ }
+}
+
+int kblock_copy(struct KBlock *b, block_idx_t src, block_idx_t dest)
+{
+ ASSERT(b);
+ ASSERT(src < b->blk_cnt);
+ ASSERT(dest < b->blk_cnt);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, src))
+ return EOF;
+
+ b->priv.curr_blk = dest;
+ kblock_setDirty(b, true);
+ return 0;
+ }
+ else if (kblock_partialWrite(b))
+ {
+ uint8_t buf[16];
+ size_t blk_size = b->blk_size;
+ size_t offset = 0;
+
+ while (blk_size)
+ {
+ size_t size = MIN(sizeof(buf), blk_size);
+ if (kblock_readDirect(b, src, buf, offset, size) != size)
+ return EOF;
+ if (kblock_writeDirect(b, dest, buf, offset, size) != size)
+ return EOF;
+
+ blk_size -= size;
+ offset += size;
+ }
+ return 0;
+ }
+ else
+ {
+ ASSERT(0);
+ return EOF;
+ }
+}
+
+int kblock_swLoad(struct KBlock *b, block_idx_t index)
+{
+ /*
+ * Since this is a low level API, the index here is a fisical index.
+ * If we call another low level API, logical to fisical translation
+ * would be applied twice.
+ * In order to avoid this we subtract the start block index.
+ */
+ ASSERT(index >= b->priv.blk_start);
+ return (kblock_readDirect(b, index - b->priv.blk_start, b->priv.buf, 0, b->blk_size) == b->blk_size) ? 0 : EOF;
+}
+
+int kblock_swStore(struct KBlock *b, block_idx_t index)
+{
+ /*
+ * Since this is a low level API, the index here is a fisical index.
+ * If we call another low level API, logical to fisical translation
+ * would be applied twice.
+ * In order to avoid this we subtract the start block index.
+ */
+ ASSERT(index >= b->priv.blk_start);
+ return (kblock_writeDirect(b, index - b->priv.blk_start, b->priv.buf, 0, b->blk_size) == b->blk_size) ? 0 : EOF;
+}
+
+size_t kblock_swReadBuf(struct KBlock *b, void *buf, size_t offset, size_t size)
+{
+ ASSERT(buf);
+ ASSERT(offset + size <= b->blk_size);
+
+ memcpy(buf, (uint8_t *)b->priv.buf + offset, size);
+ return size;
+}
+
+size_t kblock_swWriteBuf(struct KBlock *b, const void *buf, size_t offset, size_t size)
+{
+ ASSERT(buf);
+ ASSERT(offset + size <= b->blk_size);
+ memcpy((uint8_t *)b->priv.buf + offset, buf, size);
+ return size;
+}
+
+int kblock_swClose(UNUSED_ARG(struct KBlock, *b))
+{
+ return 0;
}