lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
#ifndef NVM_H
|
|
|
|
#define NVM_H
|
|
|
|
|
|
|
|
enum {
|
|
|
|
NVM_IO_OK = 0,
|
|
|
|
NVM_IO_REQUEUE = 1,
|
|
|
|
NVM_IO_DONE = 2,
|
|
|
|
NVM_IO_ERR = 3,
|
|
|
|
|
|
|
|
NVM_IOTYPE_NONE = 0,
|
|
|
|
NVM_IOTYPE_GC = 1,
|
|
|
|
};
|
|
|
|
|
|
|
|
#ifdef CONFIG_NVM
|
|
|
|
|
|
|
|
#include <linux/blkdev.h>
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/file.h>
|
|
|
|
#include <linux/dmapool.h>
|
|
|
|
|
|
|
|
enum {
|
|
|
|
/* HW Responsibilities */
|
|
|
|
NVM_RSP_L2P = 1 << 0,
|
|
|
|
NVM_RSP_ECC = 1 << 1,
|
|
|
|
|
|
|
|
/* Physical Adressing Mode */
|
|
|
|
NVM_ADDRMODE_LINEAR = 0,
|
|
|
|
NVM_ADDRMODE_CHANNEL = 1,
|
|
|
|
|
|
|
|
/* Plane programming mode for LUN */
|
|
|
|
NVM_PLANE_SINGLE = 0,
|
|
|
|
NVM_PLANE_DOUBLE = 1,
|
|
|
|
NVM_PLANE_QUAD = 2,
|
|
|
|
|
|
|
|
/* Status codes */
|
|
|
|
NVM_RSP_SUCCESS = 0x0,
|
|
|
|
NVM_RSP_NOT_CHANGEABLE = 0x1,
|
|
|
|
NVM_RSP_ERR_FAILWRITE = 0x40ff,
|
|
|
|
NVM_RSP_ERR_EMPTYPAGE = 0x42ff,
|
|
|
|
|
|
|
|
/* Device opcodes */
|
|
|
|
NVM_OP_HBREAD = 0x02,
|
|
|
|
NVM_OP_HBWRITE = 0x81,
|
|
|
|
NVM_OP_PWRITE = 0x91,
|
|
|
|
NVM_OP_PREAD = 0x92,
|
|
|
|
NVM_OP_ERASE = 0x90,
|
|
|
|
|
|
|
|
/* PPA Command Flags */
|
|
|
|
NVM_IO_SNGL_ACCESS = 0x0,
|
|
|
|
NVM_IO_DUAL_ACCESS = 0x1,
|
|
|
|
NVM_IO_QUAD_ACCESS = 0x2,
|
|
|
|
|
|
|
|
NVM_IO_SUSPEND = 0x80,
|
|
|
|
NVM_IO_SLC_MODE = 0x100,
|
|
|
|
NVM_IO_SCRAMBLE_DISABLE = 0x200,
|
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_id_group {
|
|
|
|
u8 mtype;
|
|
|
|
u8 fmtype;
|
|
|
|
u16 res16;
|
|
|
|
u8 num_ch;
|
|
|
|
u8 num_lun;
|
|
|
|
u8 num_pln;
|
|
|
|
u16 num_blk;
|
|
|
|
u16 num_pg;
|
|
|
|
u16 fpg_sz;
|
|
|
|
u16 csecs;
|
|
|
|
u16 sos;
|
|
|
|
u32 trdt;
|
|
|
|
u32 trdm;
|
|
|
|
u32 tprt;
|
|
|
|
u32 tprm;
|
|
|
|
u32 tbet;
|
|
|
|
u32 tbem;
|
|
|
|
u32 mpos;
|
|
|
|
u16 cpar;
|
|
|
|
u8 res[913];
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct nvm_addr_format {
|
|
|
|
u8 ch_offset;
|
|
|
|
u8 ch_len;
|
|
|
|
u8 lun_offset;
|
|
|
|
u8 lun_len;
|
|
|
|
u8 pln_offset;
|
|
|
|
u8 pln_len;
|
|
|
|
u8 blk_offset;
|
|
|
|
u8 blk_len;
|
|
|
|
u8 pg_offset;
|
|
|
|
u8 pg_len;
|
|
|
|
u8 sect_offset;
|
|
|
|
u8 sect_len;
|
|
|
|
u8 res[4];
|
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_id {
|
|
|
|
u8 ver_id;
|
|
|
|
u8 vmnt;
|
|
|
|
u8 cgrps;
|
|
|
|
u8 res[5];
|
|
|
|
u32 cap;
|
|
|
|
u32 dom;
|
|
|
|
struct nvm_addr_format ppaf;
|
|
|
|
u8 ppat;
|
|
|
|
u8 resv[224];
|
|
|
|
struct nvm_id_group groups[4];
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct nvm_target {
|
|
|
|
struct list_head list;
|
|
|
|
struct nvm_tgt_type *type;
|
|
|
|
struct gendisk *disk;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_tgt_instance {
|
|
|
|
struct nvm_tgt_type *tt;
|
|
|
|
};
|
|
|
|
|
|
|
|
#define ADDR_EMPTY (~0ULL)
|
|
|
|
|
|
|
|
#define NVM_VERSION_MAJOR 1
|
|
|
|
#define NVM_VERSION_MINOR 0
|
|
|
|
#define NVM_VERSION_PATCH 0
|
|
|
|
|
|
|
|
#define NVM_SEC_BITS (8)
|
|
|
|
#define NVM_PL_BITS (6)
|
|
|
|
#define NVM_PG_BITS (16)
|
|
|
|
#define NVM_BLK_BITS (16)
|
|
|
|
#define NVM_LUN_BITS (10)
|
|
|
|
#define NVM_CH_BITS (8)
|
|
|
|
|
|
|
|
struct ppa_addr {
|
|
|
|
union {
|
|
|
|
/* Channel-based PPA format in nand 4x2x2x2x8x10 */
|
|
|
|
struct {
|
2015-11-02 19:12:27 +03:00
|
|
|
u64 ch : 4;
|
|
|
|
u64 sec : 2; /* 4 sectors per page */
|
|
|
|
u64 pl : 2; /* 4 planes per LUN */
|
|
|
|
u64 lun : 2; /* 4 LUNs per channel */
|
|
|
|
u64 pg : 8; /* 256 pages per block */
|
|
|
|
u64 blk : 10;/* 1024 blocks per plane */
|
|
|
|
u64 resved : 36;
|
lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
} chnl;
|
|
|
|
|
|
|
|
/* Generic structure for all addresses */
|
|
|
|
struct {
|
2015-11-02 19:12:27 +03:00
|
|
|
u64 sec : NVM_SEC_BITS;
|
|
|
|
u64 pl : NVM_PL_BITS;
|
|
|
|
u64 pg : NVM_PG_BITS;
|
|
|
|
u64 blk : NVM_BLK_BITS;
|
|
|
|
u64 lun : NVM_LUN_BITS;
|
|
|
|
u64 ch : NVM_CH_BITS;
|
lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
} g;
|
|
|
|
|
2015-11-02 19:12:27 +03:00
|
|
|
u64 ppa;
|
lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
};
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct nvm_rq {
|
|
|
|
struct nvm_tgt_instance *ins;
|
|
|
|
struct nvm_dev *dev;
|
|
|
|
|
|
|
|
struct bio *bio;
|
|
|
|
|
|
|
|
union {
|
|
|
|
struct ppa_addr ppa_addr;
|
|
|
|
dma_addr_t dma_ppa_list;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct ppa_addr *ppa_list;
|
|
|
|
|
|
|
|
void *metadata;
|
|
|
|
dma_addr_t dma_metadata;
|
|
|
|
|
|
|
|
uint8_t opcode;
|
|
|
|
uint16_t nr_pages;
|
|
|
|
uint16_t flags;
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline struct nvm_rq *nvm_rq_from_pdu(void *pdu)
|
|
|
|
{
|
|
|
|
return pdu - sizeof(struct nvm_rq);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void *nvm_rq_to_pdu(struct nvm_rq *rqdata)
|
|
|
|
{
|
|
|
|
return rqdata + 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct nvm_block;
|
|
|
|
|
|
|
|
typedef int (nvm_l2p_update_fn)(u64, u32, __le64 *, void *);
|
|
|
|
typedef int (nvm_bb_update_fn)(u32, void *, unsigned int, void *);
|
|
|
|
typedef int (nvm_id_fn)(struct request_queue *, struct nvm_id *);
|
|
|
|
typedef int (nvm_get_l2p_tbl_fn)(struct request_queue *, u64, u32,
|
|
|
|
nvm_l2p_update_fn *, void *);
|
|
|
|
typedef int (nvm_op_bb_tbl_fn)(struct request_queue *, int, unsigned int,
|
|
|
|
nvm_bb_update_fn *, void *);
|
|
|
|
typedef int (nvm_op_set_bb_fn)(struct request_queue *, struct nvm_rq *, int);
|
|
|
|
typedef int (nvm_submit_io_fn)(struct request_queue *, struct nvm_rq *);
|
|
|
|
typedef int (nvm_erase_blk_fn)(struct request_queue *, struct nvm_rq *);
|
|
|
|
typedef void *(nvm_create_dma_pool_fn)(struct request_queue *, char *);
|
|
|
|
typedef void (nvm_destroy_dma_pool_fn)(void *);
|
|
|
|
typedef void *(nvm_dev_dma_alloc_fn)(struct request_queue *, void *, gfp_t,
|
|
|
|
dma_addr_t *);
|
|
|
|
typedef void (nvm_dev_dma_free_fn)(void *, void*, dma_addr_t);
|
|
|
|
|
|
|
|
struct nvm_dev_ops {
|
|
|
|
nvm_id_fn *identity;
|
|
|
|
nvm_get_l2p_tbl_fn *get_l2p_tbl;
|
|
|
|
nvm_op_bb_tbl_fn *get_bb_tbl;
|
|
|
|
nvm_op_set_bb_fn *set_bb;
|
|
|
|
|
|
|
|
nvm_submit_io_fn *submit_io;
|
|
|
|
nvm_erase_blk_fn *erase_block;
|
|
|
|
|
|
|
|
nvm_create_dma_pool_fn *create_dma_pool;
|
|
|
|
nvm_destroy_dma_pool_fn *destroy_dma_pool;
|
|
|
|
nvm_dev_dma_alloc_fn *dev_dma_alloc;
|
|
|
|
nvm_dev_dma_free_fn *dev_dma_free;
|
|
|
|
|
2015-11-16 17:34:36 +03:00
|
|
|
unsigned int max_phys_sect;
|
lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_lun {
|
|
|
|
int id;
|
|
|
|
|
|
|
|
int lun_id;
|
|
|
|
int chnl_id;
|
|
|
|
|
|
|
|
unsigned int nr_free_blocks; /* Number of unused blocks */
|
|
|
|
struct nvm_block *blocks;
|
|
|
|
|
|
|
|
spinlock_t lock;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_block {
|
|
|
|
struct list_head list;
|
|
|
|
struct nvm_lun *lun;
|
|
|
|
unsigned long id;
|
|
|
|
|
|
|
|
void *priv;
|
|
|
|
int type;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct nvm_dev {
|
|
|
|
struct nvm_dev_ops *ops;
|
|
|
|
|
|
|
|
struct list_head devices;
|
|
|
|
struct list_head online_targets;
|
|
|
|
|
|
|
|
/* Media manager */
|
|
|
|
struct nvmm_type *mt;
|
|
|
|
void *mp;
|
|
|
|
|
|
|
|
/* Device information */
|
|
|
|
int nr_chnls;
|
|
|
|
int nr_planes;
|
|
|
|
int luns_per_chnl;
|
|
|
|
int sec_per_pg; /* only sectors for a single page */
|
|
|
|
int pgs_per_blk;
|
|
|
|
int blks_per_lun;
|
|
|
|
int sec_size;
|
|
|
|
int oob_size;
|
|
|
|
int addr_mode;
|
|
|
|
struct nvm_addr_format addr_format;
|
|
|
|
|
|
|
|
/* Calculated/Cached values. These do not reflect the actual usable
|
|
|
|
* blocks at run-time.
|
|
|
|
*/
|
|
|
|
int max_rq_size;
|
|
|
|
int plane_mode; /* drive device in single, double or quad mode */
|
|
|
|
|
|
|
|
int sec_per_pl; /* all sectors across planes */
|
|
|
|
int sec_per_blk;
|
|
|
|
int sec_per_lun;
|
|
|
|
|
|
|
|
unsigned long total_pages;
|
|
|
|
unsigned long total_blocks;
|
|
|
|
int nr_luns;
|
|
|
|
unsigned max_pages_per_blk;
|
|
|
|
|
|
|
|
void *ppalist_pool;
|
|
|
|
|
|
|
|
struct nvm_id identity;
|
|
|
|
|
|
|
|
/* Backend device */
|
|
|
|
struct request_queue *q;
|
|
|
|
char name[DISK_NAME_LEN];
|
|
|
|
};
|
|
|
|
|
|
|
|
/* fallback conversion */
|
|
|
|
static struct ppa_addr __generic_to_linear_addr(struct nvm_dev *dev,
|
|
|
|
struct ppa_addr r)
|
|
|
|
{
|
|
|
|
struct ppa_addr l;
|
|
|
|
|
|
|
|
l.ppa = r.g.sec +
|
|
|
|
r.g.pg * dev->sec_per_pg +
|
|
|
|
r.g.blk * (dev->pgs_per_blk *
|
|
|
|
dev->sec_per_pg) +
|
|
|
|
r.g.lun * (dev->blks_per_lun *
|
|
|
|
dev->pgs_per_blk *
|
|
|
|
dev->sec_per_pg) +
|
|
|
|
r.g.ch * (dev->blks_per_lun *
|
|
|
|
dev->pgs_per_blk *
|
|
|
|
dev->luns_per_chnl *
|
|
|
|
dev->sec_per_pg);
|
|
|
|
|
|
|
|
return l;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* fallback conversion */
|
|
|
|
static struct ppa_addr __linear_to_generic_addr(struct nvm_dev *dev,
|
|
|
|
struct ppa_addr r)
|
|
|
|
{
|
|
|
|
struct ppa_addr l;
|
|
|
|
int secs, pgs, blks, luns;
|
|
|
|
sector_t ppa = r.ppa;
|
|
|
|
|
|
|
|
l.ppa = 0;
|
|
|
|
|
|
|
|
div_u64_rem(ppa, dev->sec_per_pg, &secs);
|
|
|
|
l.g.sec = secs;
|
|
|
|
|
|
|
|
sector_div(ppa, dev->sec_per_pg);
|
|
|
|
div_u64_rem(ppa, dev->sec_per_blk, &pgs);
|
|
|
|
l.g.pg = pgs;
|
|
|
|
|
|
|
|
sector_div(ppa, dev->pgs_per_blk);
|
|
|
|
div_u64_rem(ppa, dev->blks_per_lun, &blks);
|
|
|
|
l.g.blk = blks;
|
|
|
|
|
|
|
|
sector_div(ppa, dev->blks_per_lun);
|
|
|
|
div_u64_rem(ppa, dev->luns_per_chnl, &luns);
|
|
|
|
l.g.lun = luns;
|
|
|
|
|
|
|
|
sector_div(ppa, dev->luns_per_chnl);
|
|
|
|
l.g.ch = ppa;
|
|
|
|
|
|
|
|
return l;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct ppa_addr __generic_to_chnl_addr(struct ppa_addr r)
|
|
|
|
{
|
|
|
|
struct ppa_addr l;
|
|
|
|
|
|
|
|
l.ppa = 0;
|
|
|
|
|
|
|
|
l.chnl.sec = r.g.sec;
|
|
|
|
l.chnl.pl = r.g.pl;
|
|
|
|
l.chnl.pg = r.g.pg;
|
|
|
|
l.chnl.blk = r.g.blk;
|
|
|
|
l.chnl.lun = r.g.lun;
|
|
|
|
l.chnl.ch = r.g.ch;
|
|
|
|
|
|
|
|
return l;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct ppa_addr __chnl_to_generic_addr(struct ppa_addr r)
|
|
|
|
{
|
|
|
|
struct ppa_addr l;
|
|
|
|
|
|
|
|
l.ppa = 0;
|
|
|
|
|
|
|
|
l.g.sec = r.chnl.sec;
|
|
|
|
l.g.pl = r.chnl.pl;
|
|
|
|
l.g.pg = r.chnl.pg;
|
|
|
|
l.g.blk = r.chnl.blk;
|
|
|
|
l.g.lun = r.chnl.lun;
|
|
|
|
l.g.ch = r.chnl.ch;
|
|
|
|
|
|
|
|
return l;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct ppa_addr addr_to_generic_mode(struct nvm_dev *dev,
|
|
|
|
struct ppa_addr gppa)
|
|
|
|
{
|
|
|
|
switch (dev->addr_mode) {
|
|
|
|
case NVM_ADDRMODE_LINEAR:
|
|
|
|
return __linear_to_generic_addr(dev, gppa);
|
|
|
|
case NVM_ADDRMODE_CHANNEL:
|
|
|
|
return __chnl_to_generic_addr(gppa);
|
|
|
|
default:
|
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
return gppa;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct ppa_addr generic_to_addr_mode(struct nvm_dev *dev,
|
|
|
|
struct ppa_addr gppa)
|
|
|
|
{
|
|
|
|
switch (dev->addr_mode) {
|
|
|
|
case NVM_ADDRMODE_LINEAR:
|
|
|
|
return __generic_to_linear_addr(dev, gppa);
|
|
|
|
case NVM_ADDRMODE_CHANNEL:
|
|
|
|
return __generic_to_chnl_addr(gppa);
|
|
|
|
default:
|
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
return gppa;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int ppa_empty(struct ppa_addr ppa_addr)
|
|
|
|
{
|
|
|
|
return (ppa_addr.ppa == ADDR_EMPTY);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void ppa_set_empty(struct ppa_addr *ppa_addr)
|
|
|
|
{
|
|
|
|
ppa_addr->ppa = ADDR_EMPTY;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct ppa_addr block_to_ppa(struct nvm_dev *dev,
|
|
|
|
struct nvm_block *blk)
|
|
|
|
{
|
|
|
|
struct ppa_addr ppa;
|
|
|
|
struct nvm_lun *lun = blk->lun;
|
|
|
|
|
|
|
|
ppa.ppa = 0;
|
|
|
|
ppa.g.blk = blk->id % dev->blks_per_lun;
|
|
|
|
ppa.g.lun = lun->lun_id;
|
|
|
|
ppa.g.ch = lun->chnl_id;
|
|
|
|
|
|
|
|
return ppa;
|
|
|
|
}
|
|
|
|
|
2015-11-05 20:41:16 +03:00
|
|
|
typedef blk_qc_t (nvm_tgt_make_rq_fn)(struct request_queue *, struct bio *);
|
lightnvm: Support for Open-Channel SSDs
Open-channel SSDs are devices that share responsibilities with the host
in order to implement and maintain features that typical SSDs keep
strictly in firmware. These include (i) the Flash Translation Layer
(FTL), (ii) bad block management, and (iii) hardware units such as the
flash controller, the interface controller, and large amounts of flash
chips. In this way, Open-channels SSDs exposes direct access to their
physical flash storage, while keeping a subset of the internal features
of SSDs.
LightNVM is a specification that gives support to Open-channel SSDs
LightNVM allows the host to manage data placement, garbage collection,
and parallelism. Device specific responsibilities such as bad block
management, FTL extensions to support atomic IOs, or metadata
persistence are still handled by the device.
The implementation of LightNVM consists of two parts: core and
(multiple) targets. The core implements functionality shared across
targets. This is initialization, teardown and statistics. The targets
implement the interface that exposes physical flash to user-space
applications. Examples of such targets include key-value store,
object-store, as well as traditional block devices, which can be
application-specific.
Contributions in this patch from:
Javier Gonzalez <jg@lightnvm.io>
Dongsheng Yang <yangds.fnst@cn.fujitsu.com>
Jesper Madsen <jmad@itu.dk>
Signed-off-by: Matias Bjørling <m@bjorling.me>
Signed-off-by: Jens Axboe <axboe@fb.com>
2015-10-28 21:54:55 +03:00
|
|
|
typedef sector_t (nvm_tgt_capacity_fn)(void *);
|
|
|
|
typedef int (nvm_tgt_end_io_fn)(struct nvm_rq *, int);
|
|
|
|
typedef void *(nvm_tgt_init_fn)(struct nvm_dev *, struct gendisk *, int, int);
|
|
|
|
typedef void (nvm_tgt_exit_fn)(void *);
|
|
|
|
|
|
|
|
struct nvm_tgt_type {
|
|
|
|
const char *name;
|
|
|
|
unsigned int version[3];
|
|
|
|
|
|
|
|
/* target entry points */
|
|
|
|
nvm_tgt_make_rq_fn *make_rq;
|
|
|
|
nvm_tgt_capacity_fn *capacity;
|
|
|
|
nvm_tgt_end_io_fn *end_io;
|
|
|
|
|
|
|
|
/* module-specific init/teardown */
|
|
|
|
nvm_tgt_init_fn *init;
|
|
|
|
nvm_tgt_exit_fn *exit;
|
|
|
|
|
|
|
|
/* For internal use */
|
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
extern int nvm_register_target(struct nvm_tgt_type *);
|
|
|
|
extern void nvm_unregister_target(struct nvm_tgt_type *);
|
|
|
|
|
|
|
|
extern void *nvm_dev_dma_alloc(struct nvm_dev *, gfp_t, dma_addr_t *);
|
|
|
|
extern void nvm_dev_dma_free(struct nvm_dev *, void *, dma_addr_t);
|
|
|
|
|
|
|
|
typedef int (nvmm_register_fn)(struct nvm_dev *);
|
|
|
|
typedef void (nvmm_unregister_fn)(struct nvm_dev *);
|
|
|
|
typedef struct nvm_block *(nvmm_get_blk_fn)(struct nvm_dev *,
|
|
|
|
struct nvm_lun *, unsigned long);
|
|
|
|
typedef void (nvmm_put_blk_fn)(struct nvm_dev *, struct nvm_block *);
|
|
|
|
typedef int (nvmm_open_blk_fn)(struct nvm_dev *, struct nvm_block *);
|
|
|
|
typedef int (nvmm_close_blk_fn)(struct nvm_dev *, struct nvm_block *);
|
|
|
|
typedef void (nvmm_flush_blk_fn)(struct nvm_dev *, struct nvm_block *);
|
|
|
|
typedef int (nvmm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *);
|
|
|
|
typedef int (nvmm_end_io_fn)(struct nvm_rq *, int);
|
|
|
|
typedef int (nvmm_erase_blk_fn)(struct nvm_dev *, struct nvm_block *,
|
|
|
|
unsigned long);
|
|
|
|
typedef struct nvm_lun *(nvmm_get_lun_fn)(struct nvm_dev *, int);
|
|
|
|
typedef void (nvmm_free_blocks_print_fn)(struct nvm_dev *);
|
|
|
|
|
|
|
|
struct nvmm_type {
|
|
|
|
const char *name;
|
|
|
|
unsigned int version[3];
|
|
|
|
|
|
|
|
nvmm_register_fn *register_mgr;
|
|
|
|
nvmm_unregister_fn *unregister_mgr;
|
|
|
|
|
|
|
|
/* Block administration callbacks */
|
|
|
|
nvmm_get_blk_fn *get_blk;
|
|
|
|
nvmm_put_blk_fn *put_blk;
|
|
|
|
nvmm_open_blk_fn *open_blk;
|
|
|
|
nvmm_close_blk_fn *close_blk;
|
|
|
|
nvmm_flush_blk_fn *flush_blk;
|
|
|
|
|
|
|
|
nvmm_submit_io_fn *submit_io;
|
|
|
|
nvmm_end_io_fn *end_io;
|
|
|
|
nvmm_erase_blk_fn *erase_blk;
|
|
|
|
|
|
|
|
/* Configuration management */
|
|
|
|
nvmm_get_lun_fn *get_lun;
|
|
|
|
|
|
|
|
/* Statistics */
|
|
|
|
nvmm_free_blocks_print_fn *free_blocks_print;
|
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
extern int nvm_register_mgr(struct nvmm_type *);
|
|
|
|
extern void nvm_unregister_mgr(struct nvmm_type *);
|
|
|
|
|
|
|
|
extern struct nvm_block *nvm_get_blk(struct nvm_dev *, struct nvm_lun *,
|
|
|
|
unsigned long);
|
|
|
|
extern void nvm_put_blk(struct nvm_dev *, struct nvm_block *);
|
|
|
|
|
|
|
|
extern int nvm_register(struct request_queue *, char *,
|
|
|
|
struct nvm_dev_ops *);
|
|
|
|
extern void nvm_unregister(char *);
|
|
|
|
|
|
|
|
extern int nvm_submit_io(struct nvm_dev *, struct nvm_rq *);
|
|
|
|
extern int nvm_erase_blk(struct nvm_dev *, struct nvm_block *);
|
|
|
|
#else /* CONFIG_NVM */
|
|
|
|
struct nvm_dev_ops;
|
|
|
|
|
|
|
|
static inline int nvm_register(struct request_queue *q, char *disk_name,
|
|
|
|
struct nvm_dev_ops *ops)
|
|
|
|
{
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
static inline void nvm_unregister(char *disk_name) {}
|
|
|
|
#endif /* CONFIG_NVM */
|
|
|
|
#endif /* LIGHTNVM.H */
|