+ ASSERT(offset + size <= b->blk_size);
+ LOG_INFO("blk_idx %ld, offset %u, size %u\n", idx, offset, size);
+
+ if (kblock_buffered(b) && idx == b->priv.curr_blk)
+ return kblock_readBuf(b, buf, offset, size);
+ else
+ return kblock_readDirect(b, idx, buf, offset, size);
+}
+
+
+int kblock_flush(struct KBlock *b)
+{
+ ASSERT(b);
+
+ if (!kblock_buffered(b))
+ return 0;
+
+ if (kblock_cacheDirty(b))
+ {
+ LOG_INFO("flushing block %ld\n", b->priv.curr_blk);
+ if (kblock_store(b, b->priv.curr_blk) == 0)
+ kblock_setDirty(b, false);
+ else
+ return EOF;
+ }
+ return 0;
+}
+
+
+static bool kblock_loadPage(struct KBlock *b, block_idx_t idx)
+{
+ ASSERT(b);
+
+ if (idx != b->priv.curr_blk)
+ {
+ LOG_INFO("loading block %ld\n", idx);
+ if (kblock_flush(b) != 0 || kblock_load(b, idx) != 0)
+ return false;
+
+ b->priv.curr_blk = idx;
+ }
+ return true;
+}
+
+
+int kblock_trim(struct KBlock *b, block_idx_t start, block_idx_t count)
+{
+ ASSERT(start + count <= b->blk_cnt);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, start))
+ return EOF;
+ }
+
+ b->priv.blk_start += start;
+ b->priv.curr_blk = 0; // adjust logical address
+ b->blk_cnt = count;
+ return 0;
+}
+
+
+size_t kblock_write(struct KBlock *b, block_idx_t idx, const void *buf, size_t offset, size_t size)
+{
+ ASSERT(b);
+ ASSERT(buf);
+ ASSERT(idx < b->blk_cnt);
+ ASSERT(offset + size <= b->blk_size);
+
+ LOG_INFO("blk_idx %ld, offset %u, size %u\n", idx, offset, size);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, idx))
+ return 0;
+
+ kblock_setDirty(b, true);
+ return kblock_writeBuf(b, buf, offset, size);
+ }
+ else
+ {
+ #ifdef _DEBUG
+ if (offset != 0 || size != b->blk_size)
+ ASSERT(kblock_partialWrite(b));
+ #endif
+ return kblock_writeDirect(b, idx, buf, offset, size);
+ }
+}
+
+int kblock_copy(struct KBlock *b, block_idx_t src, block_idx_t dest)
+{
+ ASSERT(b);
+ ASSERT(src < b->blk_cnt);
+ ASSERT(dest < b->blk_cnt);
+
+ if (kblock_buffered(b))
+ {
+ if (!kblock_loadPage(b, src))
+ return EOF;
+
+ b->priv.curr_blk = dest;
+ kblock_setDirty(b, true);
+ return 0;
+ }
+ else if (kblock_partialWrite(b))
+ {
+ uint8_t buf[16];
+ size_t blk_size = b->blk_size;
+ size_t offset = 0;
+
+ while (blk_size)
+ {
+ size_t size = MIN(sizeof(buf), blk_size);
+ if (kblock_readDirect(b, src, buf, offset, size) != size)
+ return EOF;
+ if (kblock_writeDirect(b, dest, buf, offset, size) != size)
+ return EOF;
+
+ blk_size -= size;
+ offset += size;
+ }
+ return 0;
+ }
+ else
+ {
+ ASSERT(0);
+ return EOF;
+ }
+}
+
+int kblock_swLoad(struct KBlock *b, block_idx_t index)
+{
+ /*
+ * Since this is a low level API, the index here is a fisical index.
+ * If we call another low level API, logical to fisical translation
+ * would be applied twice.
+ * In order to avoid this we subtract the start block index.
+ */
+ ASSERT(index >= b->priv.blk_start);
+ return (kblock_readDirect(b, index - b->priv.blk_start, b->priv.buf, 0, b->blk_size) == b->blk_size) ? 0 : EOF;
+}
+
+int kblock_swStore(struct KBlock *b, block_idx_t index)
+{
+ /*
+ * Since this is a low level API, the index here is a fisical index.
+ * If we call another low level API, logical to fisical translation
+ * would be applied twice.
+ * In order to avoid this we subtract the start block index.
+ */
+ ASSERT(index >= b->priv.blk_start);
+ return (kblock_writeDirect(b, index - b->priv.blk_start, b->priv.buf, 0, b->blk_size) == b->blk_size) ? 0 : EOF;
+}
+
+size_t kblock_swReadBuf(struct KBlock *b, void *buf, size_t offset, size_t size)
+{
+ ASSERT(buf);
+ ASSERT(offset + size <= b->blk_size);
+
+ memcpy(buf, (uint8_t *)b->priv.buf + offset, size);
+ return size;
+}
+
+size_t kblock_swWriteBuf(struct KBlock *b, const void *buf, size_t offset, size_t size)
+{
+ ASSERT(buf);
+ ASSERT(offset + size <= b->blk_size);
+ memcpy((uint8_t *)b->priv.buf + offset, buf, size);
+ return size;
+}
+
+int kblock_swClose(UNUSED_ARG(struct KBlock, *b))
+{
+ return 0;