2 * NAND Flash Controller Device Driver
3 * Copyright (c) 2009, Intel Corporation and its suppliers.
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms and conditions of the GNU General Public License,
7 * version 2, as published by the Free Software Foundation.
9 * This program is distributed in the hope it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
14 * You should have received a copy of the GNU General Public License along with
15 * this program; if not, write to the Free Software Foundation, Inc.,
16 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21 #include <linux/slab.h>
31 #define BLK_FROM_ADDR(addr) ((u32)(addr >> DeviceInfo.nBitsInBlockDataSize))
32 #define PAGE_FROM_ADDR(addr, Block) ((u16)((addr - (u64)Block * \
33 DeviceInfo.wBlockDataSize) >> DeviceInfo.nBitsInPageDataSize))
35 #define IS_SPARE_BLOCK(blk) (BAD_BLOCK != (pbt[blk] &\
36 BAD_BLOCK) && SPARE_BLOCK == (pbt[blk] & SPARE_BLOCK))
38 #define IS_DATA_BLOCK(blk) (0 == (pbt[blk] & BAD_BLOCK))
40 #define IS_DISCARDED_BLOCK(blk) (BAD_BLOCK != (pbt[blk] &\
41 BAD_BLOCK) && DISCARD_BLOCK == (pbt[blk] & DISCARD_BLOCK))
43 #define IS_BAD_BLOCK(blk) (BAD_BLOCK == (pbt[blk] & BAD_BLOCK))
46 void debug_boundary_lineno_error(int chnl
, int limit
, int no
,
47 int lineno
, char *filename
)
50 printk(KERN_ERR
"Boundary Check Fail value %d >= limit %d, "
51 "at %s:%d. Other info:%d. Aborting...\n",
52 chnl
, limit
, filename
, lineno
, no
);
54 /* static int globalmemsize; */
57 static u16
FTL_Cache_If_Hit(u64 dwPageAddr
);
58 static int FTL_Cache_Read(u64 dwPageAddr
);
59 static void FTL_Cache_Read_Page(u8
*pData
, u64 dwPageAddr
,
61 static void FTL_Cache_Write_Page(u8
*pData
, u64 dwPageAddr
,
62 u8 cache_blk
, u16 flag
);
63 static int FTL_Cache_Write(void);
64 static void FTL_Calculate_LRU(void);
65 static u32
FTL_Get_Block_Index(u32 wBlockNum
);
67 static int FTL_Search_Block_Table_IN_Block(u32 BT_Block
,
68 u8 BT_Tag
, u16
*Page
);
69 static int FTL_Read_Block_Table(void);
70 static int FTL_Write_Block_Table(int wForce
);
71 static int FTL_Write_Block_Table_Data(void);
72 static int FTL_Check_Block_Table(int wOldTable
);
73 static int FTL_Static_Wear_Leveling(void);
74 static u32
FTL_Replace_Block_Table(void);
75 static int FTL_Write_IN_Progress_Block_Table_Page(void);
77 static u32
FTL_Get_Page_Num(u64 length
);
78 static u64
FTL_Get_Physical_Block_Addr(u64 blk_addr
);
80 static u32
FTL_Replace_OneBlock(u32 wBlockNum
,
82 static u32
FTL_Replace_LWBlock(u32 wBlockNum
,
83 int *pGarbageCollect
);
84 static u32
FTL_Replace_MWBlock(void);
85 static int FTL_Replace_Block(u64 blk_addr
);
86 static int FTL_Adjust_Relative_Erase_Count(u32 Index_of_MAX
);
88 struct device_info_tag DeviceInfo
;
89 struct flash_cache_tag Cache
;
90 static struct spectra_l2_cache_info cache_l2
;
92 static u8
*cache_l2_page_buf
;
93 static u8
*cache_l2_blk_buf
;
99 static u16 g_wBlockTableOffset
;
100 static u32 g_wBlockTableIndex
;
101 static u8 g_cBlockTableStatus
;
103 static u8
*g_pTempBuf
;
104 static u8
*flag_check_blk_table
;
105 static u8
*tmp_buf_search_bt_in_block
;
106 static u8
*spare_buf_search_bt_in_block
;
107 static u8
*spare_buf_bt_search_bt_in_block
;
108 static u8
*tmp_buf1_read_blk_table
;
109 static u8
*tmp_buf2_read_blk_table
;
110 static u8
*flags_static_wear_leveling
;
111 static u8
*tmp_buf_write_blk_table_data
;
112 static u8
*tmp_buf_read_disturbance
;
114 u8
*buf_read_page_main_spare
;
115 u8
*buf_write_page_main_spare
;
116 u8
*buf_read_page_spare
;
117 u8
*buf_get_bad_block
;
119 #if (RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE && CMD_DMA)
120 struct flash_cache_delta_list_tag int_cache
[MAX_CHANS
+ MAX_DESCS
];
121 struct flash_cache_tag cache_start_copy
;
124 int g_wNumFreeBlocks
;
128 static u8 bt_flag
= FIRST_BT_ID
;
129 static u8 bt_block_changed
;
131 static u16 cache_block_to_write
;
132 static u8 last_erased
= FIRST_BT_ID
;
135 static u8 BT_GC_Called
;
138 #define COPY_BACK_BUF_NUM 10
140 static u8 ftl_cmd_cnt
; /* Init value is 0 */
143 u8
*g_pBTStartingCopy
;
144 u8
*g_pWearCounterCopy
;
145 u16
*g_pReadCounterCopy
;
146 u8
*g_pBlockTableCopies
;
147 u8
*g_pNextBlockTable
;
148 static u8
*cp_back_buf_copies
[COPY_BACK_BUF_NUM
];
149 static int cp_back_buf_idx
;
151 static u8
*g_temp_buf
;
153 #pragma pack(push, 1)
155 struct BTableChangesDelta
{
158 u16 g_wBlockTableOffset
;
159 u32 g_wBlockTableIndex
;
170 struct BTableChangesDelta
*p_BTableChangesDelta
;
174 #define MARK_BLOCK_AS_BAD(blocknode) (blocknode |= BAD_BLOCK)
175 #define MARK_BLK_AS_DISCARD(blk) (blk = (blk & ~SPARE_BLOCK) | DISCARD_BLOCK)
177 #define FTL_Get_LBAPBA_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
179 #define FTL_Get_WearCounter_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
181 #define FTL_Get_ReadCounter_Table_Mem_Size_Bytes() (DeviceInfo.wDataBlockNum *\
183 #if SUPPORT_LARGE_BLOCKNUM
184 #define FTL_Get_LBAPBA_Table_Flash_Size_Bytes() (DeviceInfo.wDataBlockNum *\
187 #define FTL_Get_LBAPBA_Table_Flash_Size_Bytes() (DeviceInfo.wDataBlockNum *\
190 #define FTL_Get_WearCounter_Table_Flash_Size_Bytes \
191 FTL_Get_WearCounter_Table_Mem_Size_Bytes
192 #define FTL_Get_ReadCounter_Table_Flash_Size_Bytes \
193 FTL_Get_ReadCounter_Table_Mem_Size_Bytes
195 static u32
FTL_Get_Block_Table_Flash_Size_Bytes(void)
199 if (DeviceInfo
.MLCDevice
) {
200 byte_num
= FTL_Get_LBAPBA_Table_Flash_Size_Bytes() +
201 DeviceInfo
.wDataBlockNum
* sizeof(u8
) +
202 DeviceInfo
.wDataBlockNum
* sizeof(u16
);
204 byte_num
= FTL_Get_LBAPBA_Table_Flash_Size_Bytes() +
205 DeviceInfo
.wDataBlockNum
* sizeof(u8
);
208 byte_num
+= 4 * sizeof(u8
);
213 static u16
FTL_Get_Block_Table_Flash_Size_Pages(void)
215 return (u16
)FTL_Get_Page_Num(FTL_Get_Block_Table_Flash_Size_Bytes());
218 static int FTL_Copy_Block_Table_To_Flash(u8
*flashBuf
, u32 sizeToTx
,
221 u32 wBytesCopied
, blk_tbl_size
, wBytes
;
222 u32
*pbt
= (u32
*)g_pBlockTable
;
224 blk_tbl_size
= FTL_Get_LBAPBA_Table_Flash_Size_Bytes();
226 (wBytes
< sizeToTx
) && ((wBytes
+ sizeTxed
) < blk_tbl_size
);
228 #if SUPPORT_LARGE_BLOCKNUM
229 flashBuf
[wBytes
] = (u8
)(pbt
[(wBytes
+ sizeTxed
) / 3]
230 >> (((wBytes
+ sizeTxed
) % 3) ?
231 ((((wBytes
+ sizeTxed
) % 3) == 2) ? 0 : 8) : 16)) & 0xFF;
233 flashBuf
[wBytes
] = (u8
)(pbt
[(wBytes
+ sizeTxed
) / 2]
234 >> (((wBytes
+ sizeTxed
) % 2) ? 0 : 8)) & 0xFF;
238 sizeTxed
= (sizeTxed
> blk_tbl_size
) ? (sizeTxed
- blk_tbl_size
) : 0;
239 blk_tbl_size
= FTL_Get_WearCounter_Table_Flash_Size_Bytes();
240 wBytesCopied
= wBytes
;
241 wBytes
= ((blk_tbl_size
- sizeTxed
) > (sizeToTx
- wBytesCopied
)) ?
242 (sizeToTx
- wBytesCopied
) : (blk_tbl_size
- sizeTxed
);
243 memcpy(flashBuf
+ wBytesCopied
, g_pWearCounter
+ sizeTxed
, wBytes
);
245 sizeTxed
= (sizeTxed
> blk_tbl_size
) ? (sizeTxed
- blk_tbl_size
) : 0;
247 if (DeviceInfo
.MLCDevice
) {
248 blk_tbl_size
= FTL_Get_ReadCounter_Table_Flash_Size_Bytes();
249 wBytesCopied
+= wBytes
;
250 for (wBytes
= 0; ((wBytes
+ wBytesCopied
) < sizeToTx
) &&
251 ((wBytes
+ sizeTxed
) < blk_tbl_size
); wBytes
++)
252 flashBuf
[wBytes
+ wBytesCopied
] =
253 (g_pReadCounter
[(wBytes
+ sizeTxed
) / 2] >>
254 (((wBytes
+ sizeTxed
) % 2) ? 0 : 8)) & 0xFF;
257 return wBytesCopied
+ wBytes
;
260 static int FTL_Copy_Block_Table_From_Flash(u8
*flashBuf
,
261 u32 sizeToTx
, u32 sizeTxed
)
263 u32 wBytesCopied
, blk_tbl_size
, wBytes
;
264 u32
*pbt
= (u32
*)g_pBlockTable
;
266 blk_tbl_size
= FTL_Get_LBAPBA_Table_Flash_Size_Bytes();
267 for (wBytes
= 0; (wBytes
< sizeToTx
) &&
268 ((wBytes
+ sizeTxed
) < blk_tbl_size
); wBytes
++) {
269 #if SUPPORT_LARGE_BLOCKNUM
270 if (!((wBytes
+ sizeTxed
) % 3))
271 pbt
[(wBytes
+ sizeTxed
) / 3] = 0;
272 pbt
[(wBytes
+ sizeTxed
) / 3] |=
273 (flashBuf
[wBytes
] << (((wBytes
+ sizeTxed
) % 3) ?
274 ((((wBytes
+ sizeTxed
) % 3) == 2) ? 0 : 8) : 16));
276 if (!((wBytes
+ sizeTxed
) % 2))
277 pbt
[(wBytes
+ sizeTxed
) / 2] = 0;
278 pbt
[(wBytes
+ sizeTxed
) / 2] |=
279 (flashBuf
[wBytes
] << (((wBytes
+ sizeTxed
) % 2) ?
284 sizeTxed
= (sizeTxed
> blk_tbl_size
) ? (sizeTxed
- blk_tbl_size
) : 0;
285 blk_tbl_size
= FTL_Get_WearCounter_Table_Flash_Size_Bytes();
286 wBytesCopied
= wBytes
;
287 wBytes
= ((blk_tbl_size
- sizeTxed
) > (sizeToTx
- wBytesCopied
)) ?
288 (sizeToTx
- wBytesCopied
) : (blk_tbl_size
- sizeTxed
);
289 memcpy(g_pWearCounter
+ sizeTxed
, flashBuf
+ wBytesCopied
, wBytes
);
290 sizeTxed
= (sizeTxed
> blk_tbl_size
) ? (sizeTxed
- blk_tbl_size
) : 0;
292 if (DeviceInfo
.MLCDevice
) {
293 wBytesCopied
+= wBytes
;
294 blk_tbl_size
= FTL_Get_ReadCounter_Table_Flash_Size_Bytes();
295 for (wBytes
= 0; ((wBytes
+ wBytesCopied
) < sizeToTx
) &&
296 ((wBytes
+ sizeTxed
) < blk_tbl_size
); wBytes
++) {
297 if (((wBytes
+ sizeTxed
) % 2))
298 g_pReadCounter
[(wBytes
+ sizeTxed
) / 2] = 0;
299 g_pReadCounter
[(wBytes
+ sizeTxed
) / 2] |=
301 (((wBytes
+ sizeTxed
) % 2) ? 0 : 8));
305 return wBytesCopied
+wBytes
;
308 static int FTL_Insert_Block_Table_Signature(u8
*buf
, u8 tag
)
312 for (i
= 0; i
< BTSIG_BYTES
; i
++)
313 buf
[BTSIG_OFFSET
+ i
] =
314 ((tag
+ (i
* BTSIG_DELTA
) - FIRST_BT_ID
) %
315 (1 + LAST_BT_ID
-FIRST_BT_ID
)) + FIRST_BT_ID
;
320 static int FTL_Extract_Block_Table_Tag(u8
*buf
, u8
**tagarray
)
322 static u8 tag
[BTSIG_BYTES
>> 1];
323 int i
, j
, k
, tagi
, tagtemp
, status
;
325 *tagarray
= (u8
*)tag
;
328 for (i
= 0; i
< (BTSIG_BYTES
- 1); i
++) {
329 for (j
= i
+ 1; (j
< BTSIG_BYTES
) &&
330 (tagi
< (BTSIG_BYTES
>> 1)); j
++) {
331 tagtemp
= buf
[BTSIG_OFFSET
+ j
] -
332 buf
[BTSIG_OFFSET
+ i
];
333 if (tagtemp
&& !(tagtemp
% BTSIG_DELTA
)) {
334 tagtemp
= (buf
[BTSIG_OFFSET
+ i
] +
335 (1 + LAST_BT_ID
- FIRST_BT_ID
) -
337 (1 + LAST_BT_ID
- FIRST_BT_ID
);
339 for (k
= 0; k
< tagi
; k
++) {
340 if (tagtemp
== tag
[k
])
344 if (status
== FAIL
) {
345 tag
[tagi
++] = tagtemp
;
346 i
= (j
== (i
+ 1)) ? i
+ 1 : i
;
347 j
= (j
== (i
+ 1)) ? i
+ 1 : i
;
357 static int FTL_Execute_SPL_Recovery(void)
360 u32
*pbt
= (u32
*)g_pBlockTable
;
363 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
364 __FILE__
, __LINE__
, __func__
);
366 blks
= DeviceInfo
.wSpectraEndBlock
- DeviceInfo
.wSpectraStartBlock
;
367 for (j
= 0; j
<= blks
; j
++) {
369 if (((block
& BAD_BLOCK
) != BAD_BLOCK
) &&
370 ((block
& SPARE_BLOCK
) == SPARE_BLOCK
)) {
371 ret
= GLOB_LLD_Erase_Block(block
& ~BAD_BLOCK
);
373 nand_dbg_print(NAND_DBG_WARN
,
374 "NAND Program fail in %s, Line %d, "
375 "Function: %s, new Bad Block %d "
377 __FILE__
, __LINE__
, __func__
,
378 (int)(block
& ~BAD_BLOCK
));
379 MARK_BLOCK_AS_BAD(pbt
[j
]);
387 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
388 * Function: GLOB_FTL_IdentifyDevice
389 * Inputs: pointer to identify data structure
390 * Outputs: PASS / FAIL
391 * Description: the identify data structure is filled in with
392 * information for the block driver.
393 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
394 int GLOB_FTL_IdentifyDevice(struct spectra_indentfy_dev_tag
*dev_data
)
396 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
397 __FILE__
, __LINE__
, __func__
);
399 dev_data
->NumBlocks
= DeviceInfo
.wTotalBlocks
;
400 dev_data
->PagesPerBlock
= DeviceInfo
.wPagesPerBlock
;
401 dev_data
->PageDataSize
= DeviceInfo
.wPageDataSize
;
402 dev_data
->wECCBytesPerSector
= DeviceInfo
.wECCBytesPerSector
;
403 dev_data
->wDataBlockNum
= DeviceInfo
.wDataBlockNum
;
409 static int allocate_memory(void)
411 u32 block_table_size
, page_size
, block_size
, mem_size
;
418 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
419 __FILE__
, __LINE__
, __func__
);
421 page_size
= DeviceInfo
.wPageSize
;
422 block_size
= DeviceInfo
.wPagesPerBlock
* DeviceInfo
.wPageDataSize
;
424 block_table_size
= DeviceInfo
.wDataBlockNum
*
425 (sizeof(u32
) + sizeof(u8
) + sizeof(u16
));
426 block_table_size
+= (DeviceInfo
.wPageDataSize
-
427 (block_table_size
% DeviceInfo
.wPageDataSize
)) %
428 DeviceInfo
.wPageDataSize
;
430 /* Malloc memory for block tables */
431 g_pBlockTable
= kzalloc(block_table_size
, GFP_ATOMIC
);
433 goto block_table_fail
;
434 total_bytes
+= block_table_size
;
436 g_pWearCounter
= (u8
*)(g_pBlockTable
+
437 DeviceInfo
.wDataBlockNum
* sizeof(u32
));
439 if (DeviceInfo
.MLCDevice
)
440 g_pReadCounter
= (u16
*)(g_pBlockTable
+
441 DeviceInfo
.wDataBlockNum
*
442 (sizeof(u32
) + sizeof(u8
)));
444 /* Malloc memory and init for cache items */
445 for (i
= 0; i
< CACHE_ITEM_NUM
; i
++) {
446 Cache
.array
[i
].address
= NAND_CACHE_INIT_ADDR
;
447 Cache
.array
[i
].use_cnt
= 0;
448 Cache
.array
[i
].changed
= CLEAR
;
449 Cache
.array
[i
].buf
= kzalloc(Cache
.cache_item_size
,
451 if (!Cache
.array
[i
].buf
)
452 goto cache_item_fail
;
453 total_bytes
+= Cache
.cache_item_size
;
456 /* Malloc memory for IPF */
457 g_pIPF
= kzalloc(page_size
, GFP_ATOMIC
);
460 total_bytes
+= page_size
;
462 /* Malloc memory for data merging during Level2 Cache flush */
463 cache_l2_page_buf
= kmalloc(page_size
, GFP_ATOMIC
);
464 if (!cache_l2_page_buf
)
465 goto cache_l2_page_buf_fail
;
466 memset(cache_l2_page_buf
, 0xff, page_size
);
467 total_bytes
+= page_size
;
469 cache_l2_blk_buf
= kmalloc(block_size
, GFP_ATOMIC
);
470 if (!cache_l2_blk_buf
)
471 goto cache_l2_blk_buf_fail
;
472 memset(cache_l2_blk_buf
, 0xff, block_size
);
473 total_bytes
+= block_size
;
475 /* Malloc memory for temp buffer */
476 g_pTempBuf
= kzalloc(Cache
.cache_item_size
, GFP_ATOMIC
);
479 total_bytes
+= Cache
.cache_item_size
;
481 /* Malloc memory for block table blocks */
482 mem_size
= (1 + LAST_BT_ID
- FIRST_BT_ID
) * sizeof(u32
);
483 g_pBTBlocks
= kmalloc(mem_size
, GFP_ATOMIC
);
486 memset(g_pBTBlocks
, 0xff, mem_size
);
487 total_bytes
+= mem_size
;
489 /* Malloc memory for function FTL_Check_Block_Table */
490 flag_check_blk_table
= kmalloc(DeviceInfo
.wDataBlockNum
, GFP_ATOMIC
);
491 if (!flag_check_blk_table
)
492 goto flag_check_blk_table_fail
;
493 total_bytes
+= DeviceInfo
.wDataBlockNum
;
495 /* Malloc memory for function FTL_Search_Block_Table_IN_Block */
496 tmp_buf_search_bt_in_block
= kmalloc(page_size
, GFP_ATOMIC
);
497 if (!tmp_buf_search_bt_in_block
)
498 goto tmp_buf_search_bt_in_block_fail
;
499 memset(tmp_buf_search_bt_in_block
, 0xff, page_size
);
500 total_bytes
+= page_size
;
502 mem_size
= DeviceInfo
.wPageSize
- DeviceInfo
.wPageDataSize
;
503 spare_buf_search_bt_in_block
= kmalloc(mem_size
, GFP_ATOMIC
);
504 if (!spare_buf_search_bt_in_block
)
505 goto spare_buf_search_bt_in_block_fail
;
506 memset(spare_buf_search_bt_in_block
, 0xff, mem_size
);
507 total_bytes
+= mem_size
;
509 spare_buf_bt_search_bt_in_block
= kmalloc(mem_size
, GFP_ATOMIC
);
510 if (!spare_buf_bt_search_bt_in_block
)
511 goto spare_buf_bt_search_bt_in_block_fail
;
512 memset(spare_buf_bt_search_bt_in_block
, 0xff, mem_size
);
513 total_bytes
+= mem_size
;
515 /* Malloc memory for function FTL_Read_Block_Table */
516 tmp_buf1_read_blk_table
= kmalloc(page_size
, GFP_ATOMIC
);
517 if (!tmp_buf1_read_blk_table
)
518 goto tmp_buf1_read_blk_table_fail
;
519 memset(tmp_buf1_read_blk_table
, 0xff, page_size
);
520 total_bytes
+= page_size
;
522 tmp_buf2_read_blk_table
= kmalloc(page_size
, GFP_ATOMIC
);
523 if (!tmp_buf2_read_blk_table
)
524 goto tmp_buf2_read_blk_table_fail
;
525 memset(tmp_buf2_read_blk_table
, 0xff, page_size
);
526 total_bytes
+= page_size
;
528 /* Malloc memory for function FTL_Static_Wear_Leveling */
529 flags_static_wear_leveling
= kmalloc(DeviceInfo
.wDataBlockNum
,
531 if (!flags_static_wear_leveling
)
532 goto flags_static_wear_leveling_fail
;
533 total_bytes
+= DeviceInfo
.wDataBlockNum
;
535 /* Malloc memory for function FTL_Write_Block_Table_Data */
536 if (FTL_Get_Block_Table_Flash_Size_Pages() > 3)
537 mem_size
= FTL_Get_Block_Table_Flash_Size_Bytes() -
538 2 * DeviceInfo
.wPageSize
;
540 mem_size
= DeviceInfo
.wPageSize
;
541 tmp_buf_write_blk_table_data
= kmalloc(mem_size
, GFP_ATOMIC
);
542 if (!tmp_buf_write_blk_table_data
)
543 goto tmp_buf_write_blk_table_data_fail
;
544 memset(tmp_buf_write_blk_table_data
, 0xff, mem_size
);
545 total_bytes
+= mem_size
;
547 /* Malloc memory for function FTL_Read_Disturbance */
548 tmp_buf_read_disturbance
= kmalloc(block_size
, GFP_ATOMIC
);
549 if (!tmp_buf_read_disturbance
)
550 goto tmp_buf_read_disturbance_fail
;
551 memset(tmp_buf_read_disturbance
, 0xff, block_size
);
552 total_bytes
+= block_size
;
554 /* Alloc mem for function NAND_Read_Page_Main_Spare of lld_nand.c */
555 buf_read_page_main_spare
= kmalloc(DeviceInfo
.wPageSize
, GFP_ATOMIC
);
556 if (!buf_read_page_main_spare
)
557 goto buf_read_page_main_spare_fail
;
558 total_bytes
+= DeviceInfo
.wPageSize
;
560 /* Alloc mem for function NAND_Write_Page_Main_Spare of lld_nand.c */
561 buf_write_page_main_spare
= kmalloc(DeviceInfo
.wPageSize
, GFP_ATOMIC
);
562 if (!buf_write_page_main_spare
)
563 goto buf_write_page_main_spare_fail
;
564 total_bytes
+= DeviceInfo
.wPageSize
;
566 /* Alloc mem for function NAND_Read_Page_Spare of lld_nand.c */
567 buf_read_page_spare
= kmalloc(DeviceInfo
.wPageSpareSize
, GFP_ATOMIC
);
568 if (!buf_read_page_spare
)
569 goto buf_read_page_spare_fail
;
570 memset(buf_read_page_spare
, 0xff, DeviceInfo
.wPageSpareSize
);
571 total_bytes
+= DeviceInfo
.wPageSpareSize
;
573 /* Alloc mem for function NAND_Get_Bad_Block of lld_nand.c */
574 buf_get_bad_block
= kmalloc(DeviceInfo
.wPageSpareSize
, GFP_ATOMIC
);
575 if (!buf_get_bad_block
)
576 goto buf_get_bad_block_fail
;
577 memset(buf_get_bad_block
, 0xff, DeviceInfo
.wPageSpareSize
);
578 total_bytes
+= DeviceInfo
.wPageSpareSize
;
581 g_temp_buf
= kmalloc(block_size
, GFP_ATOMIC
);
584 memset(g_temp_buf
, 0xff, block_size
);
585 total_bytes
+= block_size
;
587 /* Malloc memory for copy of block table used in CDMA mode */
588 g_pBTStartingCopy
= kzalloc(block_table_size
, GFP_ATOMIC
);
589 if (!g_pBTStartingCopy
)
590 goto bt_starting_copy
;
591 total_bytes
+= block_table_size
;
593 g_pWearCounterCopy
= (u8
*)(g_pBTStartingCopy
+
594 DeviceInfo
.wDataBlockNum
* sizeof(u32
));
596 if (DeviceInfo
.MLCDevice
)
597 g_pReadCounterCopy
= (u16
*)(g_pBTStartingCopy
+
598 DeviceInfo
.wDataBlockNum
*
599 (sizeof(u32
) + sizeof(u8
)));
601 /* Malloc memory for block table copies */
602 mem_size
= 5 * DeviceInfo
.wDataBlockNum
* sizeof(u32
) +
603 5 * DeviceInfo
.wDataBlockNum
* sizeof(u8
);
604 if (DeviceInfo
.MLCDevice
)
605 mem_size
+= 5 * DeviceInfo
.wDataBlockNum
* sizeof(u16
);
606 g_pBlockTableCopies
= kzalloc(mem_size
, GFP_ATOMIC
);
607 if (!g_pBlockTableCopies
)
608 goto blk_table_copies_fail
;
609 total_bytes
+= mem_size
;
610 g_pNextBlockTable
= g_pBlockTableCopies
;
612 /* Malloc memory for Block Table Delta */
613 mem_size
= MAX_DESCS
* sizeof(struct BTableChangesDelta
);
614 g_pBTDelta
= kzalloc(mem_size
, GFP_ATOMIC
);
617 total_bytes
+= mem_size
;
618 g_pBTDelta_Free
= g_pBTDelta
;
620 /* Malloc memory for Copy Back Buffers */
621 for (j
= 0; j
< COPY_BACK_BUF_NUM
; j
++) {
622 cp_back_buf_copies
[j
] = kzalloc(block_size
, GFP_ATOMIC
);
623 if (!cp_back_buf_copies
[j
])
624 goto cp_back_buf_copies_fail
;
625 total_bytes
+= block_size
;
629 /* Malloc memory for pending commands list */
630 mem_size
= sizeof(struct pending_cmd
) * MAX_DESCS
;
631 info
.pcmds
= kzalloc(mem_size
, GFP_KERNEL
);
633 goto pending_cmds_buf_fail
;
634 total_bytes
+= mem_size
;
636 /* Malloc memory for CDMA descripter table */
637 mem_size
= sizeof(struct cdma_descriptor
) * MAX_DESCS
;
638 info
.cdma_desc_buf
= kzalloc(mem_size
, GFP_KERNEL
);
639 if (!info
.cdma_desc_buf
)
640 goto cdma_desc_buf_fail
;
641 total_bytes
+= mem_size
;
643 /* Malloc memory for Memcpy descripter table */
644 mem_size
= sizeof(struct memcpy_descriptor
) * MAX_DESCS
;
645 info
.memcp_desc_buf
= kzalloc(mem_size
, GFP_KERNEL
);
646 if (!info
.memcp_desc_buf
)
647 goto memcp_desc_buf_fail
;
648 total_bytes
+= mem_size
;
651 nand_dbg_print(NAND_DBG_WARN
,
652 "Total memory allocated in FTL layer: %d\n", total_bytes
);
658 kfree(info
.cdma_desc_buf
);
661 pending_cmds_buf_fail
:
662 cp_back_buf_copies_fail
:
665 kfree(cp_back_buf_copies
[j
]);
668 kfree(g_pBlockTableCopies
);
669 blk_table_copies_fail
:
670 kfree(g_pBTStartingCopy
);
674 kfree(buf_get_bad_block
);
677 buf_get_bad_block_fail
:
678 kfree(buf_read_page_spare
);
679 buf_read_page_spare_fail
:
680 kfree(buf_write_page_main_spare
);
681 buf_write_page_main_spare_fail
:
682 kfree(buf_read_page_main_spare
);
683 buf_read_page_main_spare_fail
:
684 kfree(tmp_buf_read_disturbance
);
685 tmp_buf_read_disturbance_fail
:
686 kfree(tmp_buf_write_blk_table_data
);
687 tmp_buf_write_blk_table_data_fail
:
688 kfree(flags_static_wear_leveling
);
689 flags_static_wear_leveling_fail
:
690 kfree(tmp_buf2_read_blk_table
);
691 tmp_buf2_read_blk_table_fail
:
692 kfree(tmp_buf1_read_blk_table
);
693 tmp_buf1_read_blk_table_fail
:
694 kfree(spare_buf_bt_search_bt_in_block
);
695 spare_buf_bt_search_bt_in_block_fail
:
696 kfree(spare_buf_search_bt_in_block
);
697 spare_buf_search_bt_in_block_fail
:
698 kfree(tmp_buf_search_bt_in_block
);
699 tmp_buf_search_bt_in_block_fail
:
700 kfree(flag_check_blk_table
);
701 flag_check_blk_table_fail
:
706 kfree(cache_l2_blk_buf
);
707 cache_l2_blk_buf_fail
:
708 kfree(cache_l2_page_buf
);
709 cache_l2_page_buf_fail
:
715 kfree(Cache
.array
[i
].buf
);
716 kfree(g_pBlockTable
);
718 printk(KERN_ERR
"Failed to kmalloc memory in %s Line %d.\n",
725 static int free_memory(void)
730 kfree(info
.memcp_desc_buf
);
731 kfree(info
.cdma_desc_buf
);
733 for (i
= COPY_BACK_BUF_NUM
- 1; i
>= 0; i
--)
734 kfree(cp_back_buf_copies
[i
]);
736 kfree(g_pBlockTableCopies
);
737 kfree(g_pBTStartingCopy
);
739 kfree(buf_get_bad_block
);
741 kfree(buf_read_page_spare
);
742 kfree(buf_write_page_main_spare
);
743 kfree(buf_read_page_main_spare
);
744 kfree(tmp_buf_read_disturbance
);
745 kfree(tmp_buf_write_blk_table_data
);
746 kfree(flags_static_wear_leveling
);
747 kfree(tmp_buf2_read_blk_table
);
748 kfree(tmp_buf1_read_blk_table
);
749 kfree(spare_buf_bt_search_bt_in_block
);
750 kfree(spare_buf_search_bt_in_block
);
751 kfree(tmp_buf_search_bt_in_block
);
752 kfree(flag_check_blk_table
);
756 for (i
= CACHE_ITEM_NUM
- 1; i
>= 0; i
--)
757 kfree(Cache
.array
[i
].buf
);
758 kfree(g_pBlockTable
);
763 static void dump_cache_l2_table(void)
766 struct spectra_l2_cache_list
*pnd
;
770 list_for_each(p
, &cache_l2
.table
.list
) {
771 pnd
= list_entry(p
, struct spectra_l2_cache_list
, list
);
772 nand_dbg_print(NAND_DBG_WARN
, "dump_cache_l2_table node: %d, logical_blk_num: %d\n", n
, pnd
->logical_blk_num
);
774 for (i = 0; i < DeviceInfo.wPagesPerBlock; i++) {
775 if (pnd->pages_array[i] != MAX_U32_VALUE)
776 nand_dbg_print(NAND_DBG_WARN, " pages_array[%d]: 0x%x\n", i, pnd->pages_array[i]);
783 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
784 * Function: GLOB_FTL_Init
786 * Outputs: PASS=0 / FAIL=1
787 * Description: allocates the memory for cache array,
788 * important data structures
789 * clears the cache array
790 * reads the block table from flash into array
791 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
792 int GLOB_FTL_Init(void)
796 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
797 __FILE__
, __LINE__
, __func__
);
799 Cache
.pages_per_item
= 1;
800 Cache
.cache_item_size
= 1 * DeviceInfo
.wPageDataSize
;
802 if (allocate_memory() != PASS
)
806 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
807 memcpy((void *)&cache_start_copy
, (void *)&Cache
,
808 sizeof(struct flash_cache_tag
));
809 memset((void *)&int_cache
, -1,
810 sizeof(struct flash_cache_delta_list_tag
) *
811 (MAX_CHANS
+ MAX_DESCS
));
816 if (FTL_Read_Block_Table() != PASS
)
819 /* Init the Level2 Cache data structure */
820 for (i
= 0; i
< BLK_NUM_FOR_L2_CACHE
; i
++)
821 cache_l2
.blk_array
[i
] = MAX_U32_VALUE
;
822 cache_l2
.cur_blk_idx
= 0;
823 cache_l2
.cur_page_num
= 0;
824 INIT_LIST_HEAD(&cache_l2
.table
.list
);
825 cache_l2
.table
.logical_blk_num
= MAX_U32_VALUE
;
827 dump_cache_l2_table();
835 static void save_blk_table_changes(u16 idx
)
838 u32
*pbt
= (u32
*)g_pBTStartingCopy
;
840 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
844 id
= idx
- MAX_CHANS
;
845 if (int_cache
[id
].item
!= -1) {
846 cache_blks
= int_cache
[id
].item
;
847 cache_start_copy
.array
[cache_blks
].address
=
848 int_cache
[id
].cache
.address
;
849 cache_start_copy
.array
[cache_blks
].changed
=
850 int_cache
[id
].cache
.changed
;
854 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
856 while (ftl_cmd
<= PendingCMD
[idx
].Tag
) {
857 if (p_BTableChangesDelta
->ValidFields
== 0x01) {
858 g_wBlockTableOffset
=
859 p_BTableChangesDelta
->g_wBlockTableOffset
;
860 } else if (p_BTableChangesDelta
->ValidFields
== 0x0C) {
861 pbt
[p_BTableChangesDelta
->BT_Index
] =
862 p_BTableChangesDelta
->BT_Entry_Value
;
863 debug_boundary_error(((
864 p_BTableChangesDelta
->BT_Index
)),
865 DeviceInfo
.wDataBlockNum
, 0);
866 } else if (p_BTableChangesDelta
->ValidFields
== 0x03) {
867 g_wBlockTableOffset
=
868 p_BTableChangesDelta
->g_wBlockTableOffset
;
870 p_BTableChangesDelta
->g_wBlockTableIndex
;
871 } else if (p_BTableChangesDelta
->ValidFields
== 0x30) {
872 g_pWearCounterCopy
[p_BTableChangesDelta
->WC_Index
] =
873 p_BTableChangesDelta
->WC_Entry_Value
;
874 } else if ((DeviceInfo
.MLCDevice
) &&
875 (p_BTableChangesDelta
->ValidFields
== 0xC0)) {
876 g_pReadCounterCopy
[p_BTableChangesDelta
->RC_Index
] =
877 p_BTableChangesDelta
->RC_Entry_Value
;
878 nand_dbg_print(NAND_DBG_DEBUG
,
879 "In event status setting read counter "
880 "GLOB_ftl_cmd_cnt %u Count %u Index %u\n",
882 p_BTableChangesDelta
->RC_Entry_Value
,
883 (unsigned int)p_BTableChangesDelta
->RC_Index
);
885 nand_dbg_print(NAND_DBG_DEBUG
,
886 "This should never occur \n");
888 p_BTableChangesDelta
+= 1;
889 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
893 static void discard_cmds(u16 n
)
895 u32
*pbt
= (u32
*)g_pBTStartingCopy
;
898 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
903 if ((PendingCMD
[n
].CMD
== WRITE_MAIN_CMD
) ||
904 (PendingCMD
[n
].CMD
== WRITE_MAIN_SPARE_CMD
)) {
905 for (k
= 0; k
< DeviceInfo
.wDataBlockNum
; k
++) {
906 if (PendingCMD
[n
].Block
== (pbt
[k
] & (~BAD_BLOCK
)))
907 MARK_BLK_AS_DISCARD(pbt
[k
]);
911 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
912 while (ftl_cmd
<= PendingCMD
[n
].Tag
) {
913 p_BTableChangesDelta
+= 1;
914 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
917 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
920 if (int_cache
[id
].item
!= -1) {
921 cache_blks
= int_cache
[id
].item
;
922 if (PendingCMD
[n
].CMD
== MEMCOPY_CMD
) {
923 if ((cache_start_copy
.array
[cache_blks
].buf
<=
924 PendingCMD
[n
].DataDestAddr
) &&
925 ((cache_start_copy
.array
[cache_blks
].buf
+
926 Cache
.cache_item_size
) >
927 PendingCMD
[n
].DataDestAddr
)) {
928 cache_start_copy
.array
[cache_blks
].address
=
929 NAND_CACHE_INIT_ADDR
;
930 cache_start_copy
.array
[cache_blks
].use_cnt
=
932 cache_start_copy
.array
[cache_blks
].changed
=
936 cache_start_copy
.array
[cache_blks
].address
=
937 int_cache
[id
].cache
.address
;
938 cache_start_copy
.array
[cache_blks
].changed
=
939 int_cache
[id
].cache
.changed
;
945 static void process_cmd_pass(int *first_failed_cmd
, u16 idx
)
947 if (0 == *first_failed_cmd
)
948 save_blk_table_changes(idx
);
953 static void process_cmd_fail_abort(int *first_failed_cmd
,
956 u32
*pbt
= (u32
*)g_pBTStartingCopy
;
959 int erase_fail
, program_fail
;
960 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
965 if (0 == *first_failed_cmd
)
966 *first_failed_cmd
= PendingCMD
[idx
].SBDCmdIndex
;
968 nand_dbg_print(NAND_DBG_DEBUG
, "Uncorrectable error has occurred "
969 "while executing %u Command %u accesing Block %u\n",
970 (unsigned int)p_BTableChangesDelta
->ftl_cmd_cnt
,
972 (unsigned int)PendingCMD
[idx
].Block
);
974 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
975 while (ftl_cmd
<= PendingCMD
[idx
].Tag
) {
976 p_BTableChangesDelta
+= 1;
977 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
980 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
981 id
= idx
- MAX_CHANS
;
983 if (int_cache
[id
].item
!= -1) {
984 cache_blks
= int_cache
[id
].item
;
985 if ((PendingCMD
[idx
].CMD
== WRITE_MAIN_CMD
)) {
986 cache_start_copy
.array
[cache_blks
].address
=
987 int_cache
[id
].cache
.address
;
988 cache_start_copy
.array
[cache_blks
].changed
= SET
;
989 } else if ((PendingCMD
[idx
].CMD
== READ_MAIN_CMD
)) {
990 cache_start_copy
.array
[cache_blks
].address
=
991 NAND_CACHE_INIT_ADDR
;
992 cache_start_copy
.array
[cache_blks
].use_cnt
= 0;
993 cache_start_copy
.array
[cache_blks
].changed
=
995 } else if (PendingCMD
[idx
].CMD
== ERASE_CMD
) {
997 } else if (PendingCMD
[idx
].CMD
== MEMCOPY_CMD
) {
1003 erase_fail
= (event
== EVENT_ERASE_FAILURE
) &&
1004 (PendingCMD
[idx
].CMD
== ERASE_CMD
);
1006 program_fail
= (event
== EVENT_PROGRAM_FAILURE
) &&
1007 ((PendingCMD
[idx
].CMD
== WRITE_MAIN_CMD
) ||
1008 (PendingCMD
[idx
].CMD
== WRITE_MAIN_SPARE_CMD
));
1010 if (erase_fail
|| program_fail
) {
1011 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
1012 if (PendingCMD
[idx
].Block
==
1013 (pbt
[i
] & (~BAD_BLOCK
)))
1014 MARK_BLOCK_AS_BAD(pbt
[i
]);
1019 static void process_cmd(int *first_failed_cmd
, u16 idx
, int event
)
1024 if (p_BTableChangesDelta
->ftl_cmd_cnt
== PendingCMD
[idx
].Tag
)
1027 if (PendingCMD
[idx
].Status
== CMD_PASS
) {
1028 process_cmd_pass(first_failed_cmd
, idx
);
1029 } else if ((PendingCMD
[idx
].Status
== CMD_FAIL
) ||
1030 (PendingCMD
[idx
].Status
== CMD_ABORT
)) {
1031 process_cmd_fail_abort(first_failed_cmd
, idx
, event
);
1032 } else if ((PendingCMD
[idx
].Status
== CMD_NOT_DONE
) &&
1033 PendingCMD
[idx
].Tag
) {
1034 nand_dbg_print(NAND_DBG_DEBUG
,
1035 " Command no. %hu is not executed\n",
1036 (unsigned int)PendingCMD
[idx
].Tag
);
1037 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
1038 while (ftl_cmd
<= PendingCMD
[idx
].Tag
) {
1039 p_BTableChangesDelta
+= 1;
1040 ftl_cmd
= p_BTableChangesDelta
->ftl_cmd_cnt
;
1046 static void process_cmd(int *first_failed_cmd
, u16 idx
, int event
)
1048 printk(KERN_ERR
"temporary workaround function. "
1049 "Should not be called! \n");
1052 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1053 * Function: GLOB_FTL_Event_Status
1055 * Outputs: Event Code
1056 * Description: It is called by SBD after hardware interrupt signalling
1057 * completion of commands chain
1058 * It does following things
1059 * get event status from LLD
1060 * analyze command chain status
1061 * determine last command executed
1063 * rebuild the block table in case of uncorrectable error
1065 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1066 int GLOB_FTL_Event_Status(int *first_failed_cmd
)
1068 int event_code
= PASS
;
1071 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1072 __FILE__
, __LINE__
, __func__
);
1074 *first_failed_cmd
= 0;
1076 event_code
= GLOB_LLD_Event_Status();
1078 switch (event_code
) {
1080 nand_dbg_print(NAND_DBG_DEBUG
, "Handling EVENT_PASS\n");
1082 case EVENT_UNCORRECTABLE_DATA_ERROR
:
1083 nand_dbg_print(NAND_DBG_DEBUG
, "Handling Uncorrectable ECC!\n");
1085 case EVENT_PROGRAM_FAILURE
:
1086 case EVENT_ERASE_FAILURE
:
1087 nand_dbg_print(NAND_DBG_WARN
, "Handling Ugly case. "
1088 "Event code: 0x%x\n", event_code
);
1089 p_BTableChangesDelta
=
1090 (struct BTableChangesDelta
*)g_pBTDelta
;
1091 for (i_P
= MAX_CHANS
; i_P
< (ftl_cmd_cnt
+ MAX_CHANS
);
1093 process_cmd(first_failed_cmd
, i_P
, event_code
);
1094 memcpy(g_pBlockTable
, g_pBTStartingCopy
,
1095 DeviceInfo
.wDataBlockNum
* sizeof(u32
));
1096 memcpy(g_pWearCounter
, g_pWearCounterCopy
,
1097 DeviceInfo
.wDataBlockNum
* sizeof(u8
));
1098 if (DeviceInfo
.MLCDevice
)
1099 memcpy(g_pReadCounter
, g_pReadCounterCopy
,
1100 DeviceInfo
.wDataBlockNum
* sizeof(u16
));
1102 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
1103 memcpy((void *)&Cache
, (void *)&cache_start_copy
,
1104 sizeof(struct flash_cache_tag
));
1105 memset((void *)&int_cache
, -1,
1106 sizeof(struct flash_cache_delta_list_tag
) *
1107 (MAX_DESCS
+ MAX_CHANS
));
1111 nand_dbg_print(NAND_DBG_WARN
,
1112 "Handling unexpected event code - 0x%x\n",
1118 memcpy(g_pBTStartingCopy
, g_pBlockTable
,
1119 DeviceInfo
.wDataBlockNum
* sizeof(u32
));
1120 memcpy(g_pWearCounterCopy
, g_pWearCounter
,
1121 DeviceInfo
.wDataBlockNum
* sizeof(u8
));
1122 if (DeviceInfo
.MLCDevice
)
1123 memcpy(g_pReadCounterCopy
, g_pReadCounter
,
1124 DeviceInfo
.wDataBlockNum
* sizeof(u16
));
1126 g_pBTDelta_Free
= g_pBTDelta
;
1128 g_pNextBlockTable
= g_pBlockTableCopies
;
1129 cp_back_buf_idx
= 0;
1131 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
1132 memcpy((void *)&cache_start_copy
, (void *)&Cache
,
1133 sizeof(struct flash_cache_tag
));
1134 memset((void *)&int_cache
, -1,
1135 sizeof(struct flash_cache_delta_list_tag
) *
1136 (MAX_DESCS
+ MAX_CHANS
));
1142 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1143 * Function: glob_ftl_execute_cmds
1146 * Description: pass thru to LLD
1147 ***************************************************************/
1148 u16
glob_ftl_execute_cmds(void)
1150 nand_dbg_print(NAND_DBG_TRACE
,
1151 "glob_ftl_execute_cmds: ftl_cmd_cnt %u\n",
1152 (unsigned int)ftl_cmd_cnt
);
1154 return glob_lld_execute_cmds();
1160 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1161 * Function: GLOB_FTL_Read Immediate
1162 * Inputs: pointer to data
1164 * Outputs: PASS / FAIL
1165 * Description: Reads one page of data into RAM directly from flash without
1166 * using or disturbing cache.It is assumed this function is called
1167 * with CMD-DMA disabled.
1168 *****************************************************************/
1169 int GLOB_FTL_Read_Immediate(u8
*read_data
, u64 addr
)
1175 u32
*pbt
= (u32
*)g_pBlockTable
;
1177 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1178 __FILE__
, __LINE__
, __func__
);
1180 Block
= BLK_FROM_ADDR(addr
);
1181 Page
= PAGE_FROM_ADDR(addr
, Block
);
1183 if (!IS_SPARE_BLOCK(Block
))
1186 phy_blk
= pbt
[Block
];
1187 wResult
= GLOB_LLD_Read_Page_Main(read_data
, phy_blk
, Page
, 1);
1189 if (DeviceInfo
.MLCDevice
) {
1190 g_pReadCounter
[phy_blk
- DeviceInfo
.wSpectraStartBlock
]++;
1191 if (g_pReadCounter
[phy_blk
- DeviceInfo
.wSpectraStartBlock
]
1192 >= MAX_READ_COUNTER
)
1193 FTL_Read_Disturbance(phy_blk
);
1194 if (g_cBlockTableStatus
!= IN_PROGRESS_BLOCK_TABLE
) {
1195 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
1196 FTL_Write_IN_Progress_Block_Table_Page();
1204 #ifdef SUPPORT_BIG_ENDIAN
1205 /*********************************************************************
1206 * Function: FTL_Invert_Block_Table
1209 * Description: Re-format the block table in ram based on BIG_ENDIAN and
1210 * LARGE_BLOCKNUM if necessary
1211 **********************************************************************/
1212 static void FTL_Invert_Block_Table(void)
1215 u32
*pbt
= (u32
*)g_pBlockTable
;
1217 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1218 __FILE__
, __LINE__
, __func__
);
1220 #ifdef SUPPORT_LARGE_BLOCKNUM
1221 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
1222 pbt
[i
] = INVERTUINT32(pbt
[i
]);
1223 g_pWearCounter
[i
] = INVERTUINT32(g_pWearCounter
[i
]);
1226 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
1227 pbt
[i
] = INVERTUINT16(pbt
[i
]);
1228 g_pWearCounter
[i
] = INVERTUINT16(g_pWearCounter
[i
]);
1234 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1235 * Function: GLOB_FTL_Flash_Init
1237 * Outputs: PASS=0 / FAIL=0x01 (based on read ID)
1238 * Description: The flash controller is initialized
1239 * The flash device is reset
1240 * Perform a flash READ ID command to confirm that a
1241 * valid device is attached and active.
1242 * The DeviceInfo structure gets filled in
1243 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1244 int GLOB_FTL_Flash_Init(void)
1248 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1249 __FILE__
, __LINE__
, __func__
);
1253 status
= GLOB_LLD_Flash_Init();
1258 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1260 * Outputs: PASS=0 / FAIL=0x01 (based on read ID)
1261 * Description: The flash controller is released
1262 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1263 int GLOB_FTL_Flash_Release(void)
1265 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1266 __FILE__
, __LINE__
, __func__
);
1268 return GLOB_LLD_Flash_Release();
1272 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1273 * Function: GLOB_FTL_Cache_Release
1276 * Description: release all allocated memory in GLOB_FTL_Init
1277 * (allocated in GLOB_FTL_Init)
1278 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1279 void GLOB_FTL_Cache_Release(void)
1281 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1282 __FILE__
, __LINE__
, __func__
);
1287 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1288 * Function: FTL_Cache_If_Hit
1289 * Inputs: Page Address
1290 * Outputs: Block number/UNHIT BLOCK
1291 * Description: Determines if the addressed page is in cache
1292 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1293 static u16
FTL_Cache_If_Hit(u64 page_addr
)
1299 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1300 __FILE__
, __LINE__
, __func__
);
1302 item
= UNHIT_CACHE_ITEM
;
1303 for (i
= 0; i
< CACHE_ITEM_NUM
; i
++) {
1304 addr
= Cache
.array
[i
].address
;
1305 if ((page_addr
>= addr
) &&
1306 (page_addr
< (addr
+ Cache
.cache_item_size
))) {
1315 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1316 * Function: FTL_Calculate_LRU
1319 * Description: Calculate the least recently block in a cache and record its
1320 * index in LRU field.
1321 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1322 static void FTL_Calculate_LRU(void)
1324 u16 i
, bCurrentLRU
, bTempCount
;
1326 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1327 __FILE__
, __LINE__
, __func__
);
1330 bTempCount
= MAX_WORD_VALUE
;
1332 for (i
= 0; i
< CACHE_ITEM_NUM
; i
++) {
1333 if (Cache
.array
[i
].use_cnt
< bTempCount
) {
1335 bTempCount
= Cache
.array
[i
].use_cnt
;
1339 Cache
.LRU
= bCurrentLRU
;
1342 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1343 * Function: FTL_Cache_Read_Page
1344 * Inputs: pointer to read buffer, logical address and cache item number
1346 * Description: Read the page from the cached block addressed by blocknumber
1347 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1348 static void FTL_Cache_Read_Page(u8
*data_buf
, u64 logic_addr
, u16 cache_item
)
1352 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1353 __FILE__
, __LINE__
, __func__
);
1355 start_addr
= Cache
.array
[cache_item
].buf
;
1356 start_addr
+= (u32
)(((logic_addr
- Cache
.array
[cache_item
].address
) >>
1357 DeviceInfo
.nBitsInPageDataSize
) * DeviceInfo
.wPageDataSize
);
1360 GLOB_LLD_MemCopy_CMD(data_buf
, start_addr
,
1361 DeviceInfo
.wPageDataSize
, 0);
1364 memcpy(data_buf
, start_addr
, DeviceInfo
.wPageDataSize
);
1367 if (Cache
.array
[cache_item
].use_cnt
< MAX_WORD_VALUE
)
1368 Cache
.array
[cache_item
].use_cnt
++;
1371 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1372 * Function: FTL_Cache_Read_All
1373 * Inputs: pointer to read buffer,block address
1374 * Outputs: PASS=0 / FAIL =1
1375 * Description: It reads pages in cache
1376 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1377 static int FTL_Cache_Read_All(u8
*pData
, u64 phy_addr
)
1384 u32
*pbt
= (u32
*)g_pBlockTable
;
1387 Block
= BLK_FROM_ADDR(phy_addr
);
1388 Page
= PAGE_FROM_ADDR(phy_addr
, Block
);
1389 PageCount
= Cache
.pages_per_item
;
1391 nand_dbg_print(NAND_DBG_DEBUG
,
1392 "%s, Line %d, Function: %s, Block: 0x%x\n",
1393 __FILE__
, __LINE__
, __func__
, Block
);
1396 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
1397 if ((pbt
[i
] & (~BAD_BLOCK
)) == Block
) {
1399 if (IS_SPARE_BLOCK(i
) || IS_BAD_BLOCK(i
) ||
1400 IS_DISCARDED_BLOCK(i
)) {
1401 /* Add by yunpeng -2008.12.3 */
1403 GLOB_LLD_MemCopy_CMD(pData
, g_temp_buf
,
1404 PageCount
* DeviceInfo
.wPageDataSize
, 0);
1408 PageCount
* DeviceInfo
.wPageDataSize
);
1412 continue; /* break ?? */
1417 if (0xffffffff == lba
)
1418 printk(KERN_ERR
"FTL_Cache_Read_All: Block is not found in BT\n");
1421 wResult
= GLOB_LLD_Read_Page_Main_cdma(pData
, Block
, Page
,
1422 PageCount
, LLD_CMD_FLAG_MODE_CDMA
);
1423 if (DeviceInfo
.MLCDevice
) {
1424 g_pReadCounter
[Block
- DeviceInfo
.wSpectraStartBlock
]++;
1425 nand_dbg_print(NAND_DBG_DEBUG
,
1426 "Read Counter modified in ftl_cmd_cnt %u"
1427 " Block %u Counter%u\n",
1428 ftl_cmd_cnt
, (unsigned int)Block
,
1429 g_pReadCounter
[Block
-
1430 DeviceInfo
.wSpectraStartBlock
]);
1432 p_BTableChangesDelta
=
1433 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
1434 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
1435 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
1436 p_BTableChangesDelta
->RC_Index
=
1437 Block
- DeviceInfo
.wSpectraStartBlock
;
1438 p_BTableChangesDelta
->RC_Entry_Value
=
1439 g_pReadCounter
[Block
- DeviceInfo
.wSpectraStartBlock
];
1440 p_BTableChangesDelta
->ValidFields
= 0xC0;
1444 if (g_pReadCounter
[Block
- DeviceInfo
.wSpectraStartBlock
] >=
1446 FTL_Read_Disturbance(Block
);
1447 if (g_cBlockTableStatus
!= IN_PROGRESS_BLOCK_TABLE
) {
1448 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
1449 FTL_Write_IN_Progress_Block_Table_Page();
1455 wResult
= GLOB_LLD_Read_Page_Main(pData
, Block
, Page
, PageCount
);
1456 if (wResult
== FAIL
)
1459 if (DeviceInfo
.MLCDevice
) {
1460 g_pReadCounter
[Block
- DeviceInfo
.wSpectraStartBlock
]++;
1461 if (g_pReadCounter
[Block
- DeviceInfo
.wSpectraStartBlock
] >=
1463 FTL_Read_Disturbance(Block
);
1464 if (g_cBlockTableStatus
!= IN_PROGRESS_BLOCK_TABLE
) {
1465 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
1466 FTL_Write_IN_Progress_Block_Table_Page();
1473 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1474 * Function: FTL_Cache_Write_All
1475 * Inputs: pointer to cache in sys memory
1476 * address of free block in flash
1477 * Outputs: PASS=0 / FAIL=1
1478 * Description: writes all the pages of the block in cache to flash
1480 * NOTE:need to make sure this works ok when cache is limited
1481 * to a partial block. This is where copy-back would be
1482 * activated. This would require knowing which pages in the
1483 * cached block are clean/dirty.Right now we only know if
1484 * the whole block is clean/dirty.
1485 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1486 static int FTL_Cache_Write_All(u8
*pData
, u64 blk_addr
)
1493 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1494 __FILE__
, __LINE__
, __func__
);
1496 nand_dbg_print(NAND_DBG_DEBUG
, "This block %d going to be written "
1497 "on %d\n", cache_block_to_write
,
1498 (u32
)(blk_addr
>> DeviceInfo
.nBitsInBlockDataSize
));
1500 Block
= BLK_FROM_ADDR(blk_addr
);
1501 Page
= PAGE_FROM_ADDR(blk_addr
, Block
);
1502 PageCount
= Cache
.pages_per_item
;
1505 if (FAIL
== GLOB_LLD_Write_Page_Main_cdma(pData
,
1506 Block
, Page
, PageCount
)) {
1507 nand_dbg_print(NAND_DBG_WARN
,
1508 "NAND Program fail in %s, Line %d, "
1509 "Function: %s, new Bad Block %d generated! "
1510 "Need Bad Block replacing.\n",
1511 __FILE__
, __LINE__
, __func__
, Block
);
1516 if (FAIL
== GLOB_LLD_Write_Page_Main(pData
, Block
, Page
, PageCount
)) {
1517 nand_dbg_print(NAND_DBG_WARN
, "NAND Program fail in %s,"
1518 " Line %d, Function %s, new Bad Block %d generated!"
1519 "Need Bad Block replacing.\n",
1520 __FILE__
, __LINE__
, __func__
, Block
);
1527 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1528 * Function: FTL_Copy_Block
1529 * Inputs: source block address
1530 * Destination block address
1531 * Outputs: PASS=0 / FAIL=1
1532 * Description: used only for static wear leveling to move the block
1533 * containing static data to new blocks(more worn)
1534 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1535 int FTL_Copy_Block(u64 old_blk_addr
, u64 blk_addr
)
1537 int i
, r1
, r2
, wResult
= PASS
;
1539 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1540 __FILE__
, __LINE__
, __func__
);
1542 for (i
= 0; i
< DeviceInfo
.wPagesPerBlock
; i
+= Cache
.pages_per_item
) {
1543 r1
= FTL_Cache_Read_All(g_pTempBuf
, old_blk_addr
+
1544 i
* DeviceInfo
.wPageDataSize
);
1545 r2
= FTL_Cache_Write_All(g_pTempBuf
, blk_addr
+
1546 i
* DeviceInfo
.wPageDataSize
);
1547 if ((ERR
== r1
) || (FAIL
== r2
)) {
1556 /* Search the block table to find out the least wear block and then return it */
1557 static u32
find_least_worn_blk_for_l2_cache(void)
1560 u32
*pbt
= (u32
*)g_pBlockTable
;
1561 u8 least_wear_cnt
= MAX_BYTE_VALUE
;
1562 u32 least_wear_blk_idx
= MAX_U32_VALUE
;
1565 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
1566 if (IS_SPARE_BLOCK(i
)) {
1567 phy_idx
= (u32
)((~BAD_BLOCK
) & pbt
[i
]);
1568 if (phy_idx
> DeviceInfo
.wSpectraEndBlock
)
1569 printk(KERN_ERR
"find_least_worn_blk_for_l2_cache: "
1570 "Too big phy block num (%d)\n", phy_idx
);
1571 if (g_pWearCounter
[phy_idx
-DeviceInfo
.wSpectraStartBlock
] < least_wear_cnt
) {
1572 least_wear_cnt
= g_pWearCounter
[phy_idx
- DeviceInfo
.wSpectraStartBlock
];
1573 least_wear_blk_idx
= i
;
1578 nand_dbg_print(NAND_DBG_WARN
,
1579 "find_least_worn_blk_for_l2_cache: "
1580 "find block %d with least worn counter (%d)\n",
1581 least_wear_blk_idx
, least_wear_cnt
);
1583 return least_wear_blk_idx
;
1588 /* Get blocks for Level2 Cache */
1589 static int get_l2_cache_blks(void)
1593 u32
*pbt
= (u32
*)g_pBlockTable
;
1595 for (n
= 0; n
< BLK_NUM_FOR_L2_CACHE
; n
++) {
1596 blk
= find_least_worn_blk_for_l2_cache();
1597 if (blk
>= DeviceInfo
.wDataBlockNum
) {
1598 nand_dbg_print(NAND_DBG_WARN
,
1599 "find_least_worn_blk_for_l2_cache: "
1600 "No enough free NAND blocks (n: %d) for L2 Cache!\n", n
);
1603 /* Tag the free block as discard in block table */
1604 pbt
[blk
] = (pbt
[blk
] & (~BAD_BLOCK
)) | DISCARD_BLOCK
;
1605 /* Add the free block to the L2 Cache block array */
1606 cache_l2
.blk_array
[n
] = pbt
[blk
] & (~BAD_BLOCK
);
1612 static int erase_l2_cache_blocks(void)
1615 u32 pblk
, lblk
= BAD_BLOCK
;
1617 u32
*pbt
= (u32
*)g_pBlockTable
;
1619 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
1620 __FILE__
, __LINE__
, __func__
);
1622 for (i
= 0; i
< BLK_NUM_FOR_L2_CACHE
; i
++) {
1623 pblk
= cache_l2
.blk_array
[i
];
1625 /* If the L2 cache block is invalid, then just skip it */
1626 if (MAX_U32_VALUE
== pblk
)
1629 BUG_ON(pblk
> DeviceInfo
.wSpectraEndBlock
);
1631 addr
= (u64
)pblk
<< DeviceInfo
.nBitsInBlockDataSize
;
1632 if (PASS
== GLOB_FTL_Block_Erase(addr
)) {
1633 /* Get logical block number of the erased block */
1634 lblk
= FTL_Get_Block_Index(pblk
);
1635 BUG_ON(BAD_BLOCK
== lblk
);
1636 /* Tag it as free in the block table */
1637 pbt
[lblk
] &= (u32
)(~DISCARD_BLOCK
);
1638 pbt
[lblk
] |= (u32
)(SPARE_BLOCK
);
1640 MARK_BLOCK_AS_BAD(pbt
[lblk
]);
1649 * Merge the valid data page in the L2 cache blocks into NAND.
1651 static int flush_l2_cache(void)
1653 struct list_head
*p
;
1654 struct spectra_l2_cache_list
*pnd
, *tmp_pnd
;
1655 u32
*pbt
= (u32
*)g_pBlockTable
;
1656 u32 phy_blk
, l2_blk
;
1661 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
1662 __FILE__
, __LINE__
, __func__
);
1664 if (list_empty(&cache_l2
.table
.list
)) /* No data to flush */
1667 //dump_cache_l2_table();
1669 if (IN_PROGRESS_BLOCK_TABLE
!= g_cBlockTableStatus
) {
1670 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
1671 FTL_Write_IN_Progress_Block_Table_Page();
1674 list_for_each(p
, &cache_l2
.table
.list
) {
1675 pnd
= list_entry(p
, struct spectra_l2_cache_list
, list
);
1676 if (IS_SPARE_BLOCK(pnd
->logical_blk_num
) ||
1677 IS_BAD_BLOCK(pnd
->logical_blk_num
) ||
1678 IS_DISCARDED_BLOCK(pnd
->logical_blk_num
)) {
1679 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d\n", __FILE__
, __LINE__
);
1680 memset(cache_l2_blk_buf
, 0xff, DeviceInfo
.wPagesPerBlock
* DeviceInfo
.wPageDataSize
);
1682 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d\n", __FILE__
, __LINE__
);
1683 phy_blk
= pbt
[pnd
->logical_blk_num
] & (~BAD_BLOCK
);
1684 ret
= GLOB_LLD_Read_Page_Main(cache_l2_blk_buf
,
1685 phy_blk
, 0, DeviceInfo
.wPagesPerBlock
);
1687 printk(KERN_ERR
"Read NAND page fail in %s, Line %d\n", __FILE__
, __LINE__
);
1691 for (i
= 0; i
< DeviceInfo
.wPagesPerBlock
; i
++) {
1692 if (pnd
->pages_array
[i
] != MAX_U32_VALUE
) {
1693 l2_blk
= cache_l2
.blk_array
[(pnd
->pages_array
[i
] >> 16) & 0xffff];
1694 l2_page
= pnd
->pages_array
[i
] & 0xffff;
1695 ret
= GLOB_LLD_Read_Page_Main(cache_l2_page_buf
, l2_blk
, l2_page
, 1);
1697 printk(KERN_ERR
"Read NAND page fail in %s, Line %d\n", __FILE__
, __LINE__
);
1699 memcpy(cache_l2_blk_buf
+ i
* DeviceInfo
.wPageDataSize
, cache_l2_page_buf
, DeviceInfo
.wPageDataSize
);
1703 /* Find a free block and tag the original block as discarded */
1704 addr
= (u64
)pnd
->logical_blk_num
<< DeviceInfo
.nBitsInBlockDataSize
;
1705 ret
= FTL_Replace_Block(addr
);
1707 printk(KERN_ERR
"FTL_Replace_Block fail in %s, Line %d\n", __FILE__
, __LINE__
);
1710 /* Write back the updated data into NAND */
1711 phy_blk
= pbt
[pnd
->logical_blk_num
] & (~BAD_BLOCK
);
1712 if (FAIL
== GLOB_LLD_Write_Page_Main(cache_l2_blk_buf
, phy_blk
, 0, DeviceInfo
.wPagesPerBlock
)) {
1713 nand_dbg_print(NAND_DBG_WARN
,
1714 "Program NAND block %d fail in %s, Line %d\n",
1715 phy_blk
, __FILE__
, __LINE__
);
1716 /* This may not be really a bad block. So just tag it as discarded. */
1717 /* Then it has a chance to be erased when garbage collection. */
1718 /* If it is really bad, then the erase will fail and it will be marked */
1719 /* as bad then. Otherwise it will be marked as free and can be used again */
1720 MARK_BLK_AS_DISCARD(pbt
[pnd
->logical_blk_num
]);
1721 /* Find another free block and write it again */
1722 FTL_Replace_Block(addr
);
1723 phy_blk
= pbt
[pnd
->logical_blk_num
] & (~BAD_BLOCK
);
1724 if (FAIL
== GLOB_LLD_Write_Page_Main(cache_l2_blk_buf
, phy_blk
, 0, DeviceInfo
.wPagesPerBlock
)) {
1725 printk(KERN_ERR
"Failed to write back block %d when flush L2 cache."
1726 "Some data will be lost!\n", phy_blk
);
1727 MARK_BLOCK_AS_BAD(pbt
[pnd
->logical_blk_num
]);
1730 /* tag the new free block as used block */
1731 pbt
[pnd
->logical_blk_num
] &= (~SPARE_BLOCK
);
1735 /* Destroy the L2 Cache table and free the memory of all nodes */
1736 list_for_each_entry_safe(pnd
, tmp_pnd
, &cache_l2
.table
.list
, list
) {
1737 list_del(&pnd
->list
);
1741 /* Erase discard L2 cache blocks */
1742 if (erase_l2_cache_blocks() != PASS
)
1743 nand_dbg_print(NAND_DBG_WARN
,
1744 " Erase L2 cache blocks error in %s, Line %d\n",
1745 __FILE__
, __LINE__
);
1747 /* Init the Level2 Cache data structure */
1748 for (i
= 0; i
< BLK_NUM_FOR_L2_CACHE
; i
++)
1749 cache_l2
.blk_array
[i
] = MAX_U32_VALUE
;
1750 cache_l2
.cur_blk_idx
= 0;
1751 cache_l2
.cur_page_num
= 0;
1752 INIT_LIST_HEAD(&cache_l2
.table
.list
);
1753 cache_l2
.table
.logical_blk_num
= MAX_U32_VALUE
;
1759 * Write back a changed victim cache item to the Level2 Cache
1760 * and update the L2 Cache table to map the change.
1761 * If the L2 Cache is full, then start to do the L2 Cache flush.
1763 static int write_back_to_l2_cache(u8
*buf
, u64 logical_addr
)
1765 u32 logical_blk_num
;
1766 u16 logical_page_num
;
1767 struct list_head
*p
;
1768 struct spectra_l2_cache_list
*pnd
, *pnd_new
;
1772 nand_dbg_print(NAND_DBG_DEBUG
, "%s, Line %d, Function: %s\n",
1773 __FILE__
, __LINE__
, __func__
);
1776 * If Level2 Cache table is empty, then it means either:
1777 * 1. This is the first time that the function called after FTL_init
1779 * 2. The Level2 Cache has just been flushed
1781 * So, 'steal' some free blocks from NAND for L2 Cache using
1782 * by just mask them as discard in the block table
1784 if (list_empty(&cache_l2
.table
.list
)) {
1785 BUG_ON(cache_l2
.cur_blk_idx
!= 0);
1786 BUG_ON(cache_l2
.cur_page_num
!= 0);
1787 BUG_ON(cache_l2
.table
.logical_blk_num
!= MAX_U32_VALUE
);
1788 if (FAIL
== get_l2_cache_blks()) {
1789 GLOB_FTL_Garbage_Collection();
1790 if (FAIL
== get_l2_cache_blks()) {
1791 printk(KERN_ALERT
"Fail to get L2 cache blks!\n");
1797 logical_blk_num
= BLK_FROM_ADDR(logical_addr
);
1798 logical_page_num
= PAGE_FROM_ADDR(logical_addr
, logical_blk_num
);
1799 BUG_ON(logical_blk_num
== MAX_U32_VALUE
);
1801 /* Write the cache item data into the current position of L2 Cache */
1807 if (FAIL
== GLOB_LLD_Write_Page_Main(buf
,
1808 cache_l2
.blk_array
[cache_l2
.cur_blk_idx
],
1809 cache_l2
.cur_page_num
, 1)) {
1810 nand_dbg_print(NAND_DBG_WARN
, "NAND Program fail in "
1811 "%s, Line %d, new Bad Block %d generated!\n",
1813 cache_l2
.blk_array
[cache_l2
.cur_blk_idx
]);
1815 /* TODO: tag the current block as bad and try again */
1822 * Update the L2 Cache table.
1824 * First seaching in the table to see whether the logical block
1825 * has been mapped. If not, then kmalloc a new node for the
1826 * logical block, fill data, and then insert it to the list.
1827 * Otherwise, just update the mapped node directly.
1830 list_for_each(p
, &cache_l2
.table
.list
) {
1831 pnd
= list_entry(p
, struct spectra_l2_cache_list
, list
);
1832 if (pnd
->logical_blk_num
== logical_blk_num
) {
1833 pnd
->pages_array
[logical_page_num
] =
1834 (cache_l2
.cur_blk_idx
<< 16) |
1835 cache_l2
.cur_page_num
;
1840 if (!found
) { /* Create new node for the logical block here */
1842 /* The logical pages to physical pages map array is
1843 * located at the end of struct spectra_l2_cache_list.
1845 node_size
= sizeof(struct spectra_l2_cache_list
) +
1846 sizeof(u32
) * DeviceInfo
.wPagesPerBlock
;
1847 pnd_new
= kmalloc(node_size
, GFP_ATOMIC
);
1849 printk(KERN_ERR
"Failed to kmalloc in %s Line %d\n",
1850 __FILE__
, __LINE__
);
1852 * TODO: Need to flush all the L2 cache into NAND ASAP
1853 * since no memory available here
1856 pnd_new
->logical_blk_num
= logical_blk_num
;
1857 for (i
= 0; i
< DeviceInfo
.wPagesPerBlock
; i
++)
1858 pnd_new
->pages_array
[i
] = MAX_U32_VALUE
;
1859 pnd_new
->pages_array
[logical_page_num
] =
1860 (cache_l2
.cur_blk_idx
<< 16) | cache_l2
.cur_page_num
;
1861 list_add(&pnd_new
->list
, &cache_l2
.table
.list
);
1864 /* Increasing the current position pointer of the L2 Cache */
1865 cache_l2
.cur_page_num
++;
1866 if (cache_l2
.cur_page_num
>= DeviceInfo
.wPagesPerBlock
) {
1867 cache_l2
.cur_blk_idx
++;
1868 if (cache_l2
.cur_blk_idx
>= BLK_NUM_FOR_L2_CACHE
) {
1869 /* The L2 Cache is full. Need to flush it now */
1870 nand_dbg_print(NAND_DBG_WARN
,
1871 "L2 Cache is full, will start to flush it\n");
1874 cache_l2
.cur_page_num
= 0;
1882 * Search in the Level2 Cache table to find the cache item.
1883 * If find, read the data from the NAND page of L2 Cache,
1884 * Otherwise, return FAIL.
1886 static int search_l2_cache(u8
*buf
, u64 logical_addr
)
1888 u32 logical_blk_num
;
1889 u16 logical_page_num
;
1890 struct list_head
*p
;
1891 struct spectra_l2_cache_list
*pnd
;
1892 u32 tmp
= MAX_U32_VALUE
;
1897 logical_blk_num
= BLK_FROM_ADDR(logical_addr
);
1898 logical_page_num
= PAGE_FROM_ADDR(logical_addr
, logical_blk_num
);
1900 list_for_each(p
, &cache_l2
.table
.list
) {
1901 pnd
= list_entry(p
, struct spectra_l2_cache_list
, list
);
1902 if (pnd
->logical_blk_num
== logical_blk_num
) {
1903 tmp
= pnd
->pages_array
[logical_page_num
];
1908 if (tmp
!= MAX_U32_VALUE
) { /* Found valid map */
1909 phy_blk
= cache_l2
.blk_array
[(tmp
>> 16) & 0xFFFF];
1910 phy_page
= tmp
& 0xFFFF;
1914 ret
= GLOB_LLD_Read_Page_Main(buf
, phy_blk
, phy_page
, 1);
1921 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1922 * Function: FTL_Cache_Write_Page
1923 * Inputs: Pointer to buffer, page address, cache block number
1924 * Outputs: PASS=0 / FAIL=1
1925 * Description: It writes the data in Cache Block
1926 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1927 static void FTL_Cache_Write_Page(u8
*pData
, u64 page_addr
,
1928 u8 cache_blk
, u16 flag
)
1933 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1934 __FILE__
, __LINE__
, __func__
);
1936 addr
= Cache
.array
[cache_blk
].address
;
1937 pDest
= Cache
.array
[cache_blk
].buf
;
1939 pDest
+= (unsigned long)(page_addr
- addr
);
1940 Cache
.array
[cache_blk
].changed
= SET
;
1942 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
1943 int_cache
[ftl_cmd_cnt
].item
= cache_blk
;
1944 int_cache
[ftl_cmd_cnt
].cache
.address
=
1945 Cache
.array
[cache_blk
].address
;
1946 int_cache
[ftl_cmd_cnt
].cache
.changed
=
1947 Cache
.array
[cache_blk
].changed
;
1949 GLOB_LLD_MemCopy_CMD(pDest
, pData
, DeviceInfo
.wPageDataSize
, flag
);
1952 memcpy(pDest
, pData
, DeviceInfo
.wPageDataSize
);
1954 if (Cache
.array
[cache_blk
].use_cnt
< MAX_WORD_VALUE
)
1955 Cache
.array
[cache_blk
].use_cnt
++;
1958 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
1959 * Function: FTL_Cache_Write
1961 * Outputs: PASS=0 / FAIL=1
1962 * Description: It writes least frequently used Cache block to flash if it
1964 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
1965 static int FTL_Cache_Write(void)
1967 int i
, bResult
= PASS
;
1968 u16 bNO
, least_count
= 0xFFFF;
1970 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
1971 __FILE__
, __LINE__
, __func__
);
1973 FTL_Calculate_LRU();
1976 nand_dbg_print(NAND_DBG_DEBUG
, "FTL_Cache_Write: "
1977 "Least used cache block is %d\n", bNO
);
1979 if (Cache
.array
[bNO
].changed
!= SET
)
1982 nand_dbg_print(NAND_DBG_DEBUG
, "FTL_Cache_Write: Cache"
1983 " Block %d containing logical block %d is dirty\n",
1985 (u32
)(Cache
.array
[bNO
].address
>>
1986 DeviceInfo
.nBitsInBlockDataSize
));
1988 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
1989 int_cache
[ftl_cmd_cnt
].item
= bNO
;
1990 int_cache
[ftl_cmd_cnt
].cache
.address
=
1991 Cache
.array
[bNO
].address
;
1992 int_cache
[ftl_cmd_cnt
].cache
.changed
= CLEAR
;
1995 bResult
= write_back_to_l2_cache(Cache
.array
[bNO
].buf
,
1996 Cache
.array
[bNO
].address
);
1998 Cache
.array
[bNO
].changed
= CLEAR
;
2000 least_count
= Cache
.array
[bNO
].use_cnt
;
2002 for (i
= 0; i
< CACHE_ITEM_NUM
; i
++) {
2005 if (Cache
.array
[i
].use_cnt
> 0)
2006 Cache
.array
[i
].use_cnt
-= least_count
;
2012 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2013 * Function: FTL_Cache_Read
2014 * Inputs: Page address
2015 * Outputs: PASS=0 / FAIL=1
2016 * Description: It reads the block from device in Cache Block
2017 * Set the LRU count to 1
2018 * Mark the Cache Block as clean
2019 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2020 static int FTL_Cache_Read(u64 logical_addr
)
2022 u64 item_addr
, phy_addr
;
2026 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2027 __FILE__
, __LINE__
, __func__
);
2029 num
= Cache
.LRU
; /* The LRU cache item will be overwritten */
2031 item_addr
= (u64
)GLOB_u64_Div(logical_addr
, Cache
.cache_item_size
) *
2032 Cache
.cache_item_size
;
2033 Cache
.array
[num
].address
= item_addr
;
2034 Cache
.array
[num
].use_cnt
= 1;
2035 Cache
.array
[num
].changed
= CLEAR
;
2038 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
2039 int_cache
[ftl_cmd_cnt
].item
= num
;
2040 int_cache
[ftl_cmd_cnt
].cache
.address
=
2041 Cache
.array
[num
].address
;
2042 int_cache
[ftl_cmd_cnt
].cache
.changed
=
2043 Cache
.array
[num
].changed
;
2047 * Search in L2 Cache. If hit, fill data into L1 Cache item buffer,
2048 * Otherwise, read it from NAND
2050 ret
= search_l2_cache(Cache
.array
[num
].buf
, logical_addr
);
2051 if (PASS
== ret
) /* Hit in L2 Cache */
2054 /* Compute the physical start address of NAND device according to */
2055 /* the logical start address of the cache item (LRU cache item) */
2056 phy_addr
= FTL_Get_Physical_Block_Addr(item_addr
) +
2057 GLOB_u64_Remainder(item_addr
, 2);
2059 return FTL_Cache_Read_All(Cache
.array
[num
].buf
, phy_addr
);
2062 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2063 * Function: FTL_Check_Block_Table
2065 * Outputs: PASS=0 / FAIL=1
2066 * Description: It checks the correctness of each block table entry
2067 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2068 static int FTL_Check_Block_Table(int wOldTable
)
2073 u32
*pbt
= (u32
*)g_pBlockTable
;
2074 u8
*pFlag
= flag_check_blk_table
;
2076 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2077 __FILE__
, __LINE__
, __func__
);
2079 if (NULL
!= pFlag
) {
2080 memset(pFlag
, FAIL
, DeviceInfo
.wDataBlockNum
);
2081 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
2082 blk_idx
= (u32
)(pbt
[i
] & (~BAD_BLOCK
));
2085 * 20081006/KBV - Changed to pFlag[i] reference
2086 * to avoid buffer overflow
2090 * 2008-10-20 Yunpeng Note: This change avoid
2091 * buffer overflow, but changed function of
2092 * the code, so it should be re-write later
2094 if ((blk_idx
> DeviceInfo
.wSpectraEndBlock
) ||
2108 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2109 * Function: FTL_Write_Block_Table
2111 * Outputs: 0=Block Table was updated. No write done. 1=Block write needs to
2113 * Description: It writes the block table
2114 * Block table always mapped to LBA 0 which inturn mapped
2115 * to any physical block
2116 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2117 static int FTL_Write_Block_Table(int wForce
)
2119 u32
*pbt
= (u32
*)g_pBlockTable
;
2120 int wSuccess
= PASS
;
2121 u32 wTempBlockTableIndex
;
2122 u16 bt_pages
, new_bt_offset
;
2123 u8 blockchangeoccured
= 0;
2125 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2126 __FILE__
, __LINE__
, __func__
);
2128 bt_pages
= FTL_Get_Block_Table_Flash_Size_Pages();
2130 if (IN_PROGRESS_BLOCK_TABLE
!= g_cBlockTableStatus
)
2133 if (PASS
== wForce
) {
2134 g_wBlockTableOffset
=
2135 (u16
)(DeviceInfo
.wPagesPerBlock
- bt_pages
);
2137 p_BTableChangesDelta
=
2138 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
2139 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
2141 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
2142 p_BTableChangesDelta
->g_wBlockTableOffset
=
2143 g_wBlockTableOffset
;
2144 p_BTableChangesDelta
->ValidFields
= 0x01;
2148 nand_dbg_print(NAND_DBG_DEBUG
,
2149 "Inside FTL_Write_Block_Table: block %d Page:%d\n",
2150 g_wBlockTableIndex
, g_wBlockTableOffset
);
2153 new_bt_offset
= g_wBlockTableOffset
+ bt_pages
+ 1;
2154 if ((0 == (new_bt_offset
% DeviceInfo
.wPagesPerBlock
)) ||
2155 (new_bt_offset
> DeviceInfo
.wPagesPerBlock
) ||
2156 (FAIL
== wSuccess
)) {
2157 wTempBlockTableIndex
= FTL_Replace_Block_Table();
2158 if (BAD_BLOCK
== wTempBlockTableIndex
)
2160 if (!blockchangeoccured
) {
2161 bt_block_changed
= 1;
2162 blockchangeoccured
= 1;
2165 g_wBlockTableIndex
= wTempBlockTableIndex
;
2166 g_wBlockTableOffset
= 0;
2167 pbt
[BLOCK_TABLE_INDEX
] = g_wBlockTableIndex
;
2169 p_BTableChangesDelta
=
2170 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
2171 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
2173 p_BTableChangesDelta
->ftl_cmd_cnt
=
2175 p_BTableChangesDelta
->g_wBlockTableOffset
=
2176 g_wBlockTableOffset
;
2177 p_BTableChangesDelta
->g_wBlockTableIndex
=
2179 p_BTableChangesDelta
->ValidFields
= 0x03;
2181 p_BTableChangesDelta
=
2182 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
2184 sizeof(struct BTableChangesDelta
);
2186 p_BTableChangesDelta
->ftl_cmd_cnt
=
2188 p_BTableChangesDelta
->BT_Index
=
2190 p_BTableChangesDelta
->BT_Entry_Value
=
2191 pbt
[BLOCK_TABLE_INDEX
];
2192 p_BTableChangesDelta
->ValidFields
= 0x0C;
2196 wSuccess
= FTL_Write_Block_Table_Data();
2197 if (FAIL
== wSuccess
)
2198 MARK_BLOCK_AS_BAD(pbt
[BLOCK_TABLE_INDEX
]);
2199 } while (FAIL
== wSuccess
);
2201 g_cBlockTableStatus
= CURRENT_BLOCK_TABLE
;
2206 static int force_format_nand(void)
2210 /* Force erase the whole unprotected physical partiton of NAND */
2211 printk(KERN_ALERT
"Start to force erase whole NAND device ...\n");
2212 printk(KERN_ALERT
"From phyical block %d to %d\n",
2213 DeviceInfo
.wSpectraStartBlock
, DeviceInfo
.wSpectraEndBlock
);
2214 for (i
= DeviceInfo
.wSpectraStartBlock
; i
<= DeviceInfo
.wSpectraEndBlock
; i
++) {
2215 if (GLOB_LLD_Erase_Block(i
))
2216 printk(KERN_ERR
"Failed to force erase NAND block %d\n", i
);
2218 printk(KERN_ALERT
"Force Erase ends. Please reboot the system ...\n");
2224 int GLOB_FTL_Flash_Format(void)
2226 //return FTL_Format_Flash(1);
2227 return force_format_nand();
2231 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2232 * Function: FTL_Search_Block_Table_IN_Block
2233 * Inputs: Block Number
2235 * Outputs: PASS / FAIL
2236 * Page contatining the block table
2237 * Description: It searches the block table in the block
2238 * passed as an argument.
2240 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2241 static int FTL_Search_Block_Table_IN_Block(u32 BT_Block
,
2242 u8 BT_Tag
, u16
*Page
)
2249 u8
*tempbuf
= tmp_buf_search_bt_in_block
;
2250 u8
*pSpareBuf
= spare_buf_search_bt_in_block
;
2251 u8
*pSpareBufBTLastPage
= spare_buf_bt_search_bt_in_block
;
2252 u8 bt_flag_last_page
= 0xFF;
2253 u8 search_in_previous_pages
= 0;
2256 nand_dbg_print(NAND_DBG_DEBUG
, "%s, Line %d, Function: %s\n",
2257 __FILE__
, __LINE__
, __func__
);
2259 nand_dbg_print(NAND_DBG_DEBUG
,
2260 "Searching block table in %u block\n",
2261 (unsigned int)BT_Block
);
2263 bt_pages
= FTL_Get_Block_Table_Flash_Size_Pages();
2265 for (i
= bt_pages
; i
< DeviceInfo
.wPagesPerBlock
;
2266 i
+= (bt_pages
+ 1)) {
2267 nand_dbg_print(NAND_DBG_DEBUG
,
2268 "Searching last IPF: %d\n", i
);
2269 Result
= GLOB_LLD_Read_Page_Main_Polling(tempbuf
,
2272 if (0 == memcmp(tempbuf
, g_pIPF
, DeviceInfo
.wPageDataSize
)) {
2273 if ((i
+ bt_pages
+ 1) < DeviceInfo
.wPagesPerBlock
) {
2276 search_in_previous_pages
= 1;
2281 if (!search_in_previous_pages
) {
2282 if (i
!= bt_pages
) {
2283 i
-= (bt_pages
+ 1);
2291 if (!search_in_previous_pages
) {
2293 nand_dbg_print(NAND_DBG_DEBUG
,
2294 "Reading the spare area of Block %u Page %u",
2295 (unsigned int)BT_Block
, i
);
2296 Result
= GLOB_LLD_Read_Page_Spare(pSpareBuf
,
2298 nand_dbg_print(NAND_DBG_DEBUG
,
2299 "Reading the spare area of Block %u Page %u",
2300 (unsigned int)BT_Block
, i
+ bt_pages
- 1);
2301 Result
= GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage
,
2302 BT_Block
, i
+ bt_pages
- 1, 1);
2305 j
= FTL_Extract_Block_Table_Tag(pSpareBuf
, &tagarray
);
2307 for (; k
< j
; k
++) {
2308 if (tagarray
[k
] == BT_Tag
)
2314 bt_flag
= tagarray
[k
];
2318 if (Result
== PASS
) {
2320 j
= FTL_Extract_Block_Table_Tag(
2321 pSpareBufBTLastPage
, &tagarray
);
2323 for (; k
< j
; k
++) {
2324 if (tagarray
[k
] == BT_Tag
)
2330 bt_flag_last_page
= tagarray
[k
];
2334 if (Result
== PASS
) {
2335 if (bt_flag
== bt_flag_last_page
) {
2336 nand_dbg_print(NAND_DBG_DEBUG
,
2337 "Block table is found"
2338 " in page after IPF "
2344 g_cBlockTableStatus
=
2345 CURRENT_BLOCK_TABLE
;
2354 if (search_in_previous_pages
)
2357 i
= i
- (bt_pages
+ 1);
2361 nand_dbg_print(NAND_DBG_DEBUG
,
2362 "Reading the spare area of Block %d Page %d",
2365 Result
= GLOB_LLD_Read_Page_Spare(pSpareBuf
, BT_Block
, i
, 1);
2366 nand_dbg_print(NAND_DBG_DEBUG
,
2367 "Reading the spare area of Block %u Page %u",
2368 (unsigned int)BT_Block
, i
+ bt_pages
- 1);
2370 Result
= GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage
,
2371 BT_Block
, i
+ bt_pages
- 1, 1);
2374 j
= FTL_Extract_Block_Table_Tag(pSpareBuf
, &tagarray
);
2376 for (; k
< j
; k
++) {
2377 if (tagarray
[k
] == BT_Tag
)
2383 bt_flag
= tagarray
[k
];
2387 if (Result
== PASS
) {
2389 j
= FTL_Extract_Block_Table_Tag(pSpareBufBTLastPage
,
2392 for (; k
< j
; k
++) {
2393 if (tagarray
[k
] == BT_Tag
)
2399 bt_flag_last_page
= tagarray
[k
];
2405 if (Result
== PASS
) {
2406 if (bt_flag
== bt_flag_last_page
) {
2407 nand_dbg_print(NAND_DBG_DEBUG
,
2408 "Block table is found "
2409 "in page prior to IPF "
2410 "at block %u page %d\n",
2411 (unsigned int)BT_Block
, i
);
2414 g_cBlockTableStatus
=
2415 IN_PROGRESS_BLOCK_TABLE
;
2425 if (Result
== FAIL
) {
2426 if ((Last_IPF
> bt_pages
) && (i
< Last_IPF
) && (!BT_Found
)) {
2428 *Page
= i
- (bt_pages
+ 1);
2430 if ((Last_IPF
== bt_pages
) && (i
< Last_IPF
) && (!BT_Found
))
2434 if (Last_IPF
== 0) {
2437 nand_dbg_print(NAND_DBG_DEBUG
, "Reading the spare area of "
2438 "Block %u Page %u", (unsigned int)BT_Block
, i
);
2440 Result
= GLOB_LLD_Read_Page_Spare(pSpareBuf
, BT_Block
, i
, 1);
2441 nand_dbg_print(NAND_DBG_DEBUG
,
2442 "Reading the spare area of Block %u Page %u",
2443 (unsigned int)BT_Block
, i
+ bt_pages
- 1);
2444 Result
= GLOB_LLD_Read_Page_Spare(pSpareBufBTLastPage
,
2445 BT_Block
, i
+ bt_pages
- 1, 1);
2448 j
= FTL_Extract_Block_Table_Tag(pSpareBuf
, &tagarray
);
2450 for (; k
< j
; k
++) {
2451 if (tagarray
[k
] == BT_Tag
)
2457 bt_flag
= tagarray
[k
];
2461 if (Result
== PASS
) {
2463 j
= FTL_Extract_Block_Table_Tag(pSpareBufBTLastPage
,
2466 for (; k
< j
; k
++) {
2467 if (tagarray
[k
] == BT_Tag
)
2473 bt_flag_last_page
= tagarray
[k
];
2477 if (Result
== PASS
) {
2478 if (bt_flag
== bt_flag_last_page
) {
2479 nand_dbg_print(NAND_DBG_DEBUG
,
2480 "Block table is found "
2481 "in page after IPF at "
2482 "block %u page %u\n",
2483 (unsigned int)BT_Block
,
2487 g_cBlockTableStatus
=
2488 CURRENT_BLOCK_TABLE
;
2503 u8
*get_blk_table_start_addr(void)
2505 return g_pBlockTable
;
2508 unsigned long get_blk_table_len(void)
2510 return DeviceInfo
.wDataBlockNum
* sizeof(u32
);
2513 u8
*get_wear_leveling_table_start_addr(void)
2515 return g_pWearCounter
;
2518 unsigned long get_wear_leveling_table_len(void)
2520 return DeviceInfo
.wDataBlockNum
* sizeof(u8
);
2523 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2524 * Function: FTL_Read_Block_Table
2526 * Outputs: PASS / FAIL
2527 * Description: read the flash spare area and find a block containing the
2528 * most recent block table(having largest block_table_counter).
2529 * Find the last written Block table in this block.
2530 * Check the correctness of Block Table
2531 * If CDMA is enabled, this function is called in
2533 * We don't need to store changes in Block table in this
2534 * function as it is called only at initialization
2536 * Note: Currently this function is called at initialization
2537 * before any read/erase/write command issued to flash so,
2538 * there is no need to wait for CDMA list to complete as of now
2539 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2540 static int FTL_Read_Block_Table(void)
2544 u8
*tempBuf
, *tagarray
;
2547 u8 block_table_found
= 0;
2553 int wBytesCopied
= 0, tempvar
;
2555 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2556 __FILE__
, __LINE__
, __func__
);
2558 tempBuf
= tmp_buf1_read_blk_table
;
2559 bt_pages
= FTL_Get_Block_Table_Flash_Size_Pages();
2561 for (j
= DeviceInfo
.wSpectraStartBlock
;
2562 j
<= (int)DeviceInfo
.wSpectraEndBlock
;
2564 status
= GLOB_LLD_Read_Page_Spare(tempBuf
, j
, 0, 1);
2566 i
= FTL_Extract_Block_Table_Tag(tempBuf
, &tagarray
);
2568 status
= GLOB_LLD_Read_Page_Main_Polling(tempBuf
,
2570 for (; k
< i
; k
++) {
2571 if (tagarray
[k
] == tempBuf
[3])
2581 nand_dbg_print(NAND_DBG_DEBUG
,
2582 "Block table is contained in Block %d %d\n",
2583 (unsigned int)j
, (unsigned int)k
);
2585 if (g_pBTBlocks
[k
-FIRST_BT_ID
] == BTBLOCK_INVAL
) {
2586 g_pBTBlocks
[k
-FIRST_BT_ID
] = j
;
2587 block_table_found
= 1;
2589 printk(KERN_ERR
"FTL_Read_Block_Table -"
2590 "This should never happens. "
2591 "Two block table have same counter %u!\n", k
);
2595 if (block_table_found
) {
2596 if (g_pBTBlocks
[FIRST_BT_ID
- FIRST_BT_ID
] != BTBLOCK_INVAL
&&
2597 g_pBTBlocks
[LAST_BT_ID
- FIRST_BT_ID
] != BTBLOCK_INVAL
) {
2599 while ((j
> FIRST_BT_ID
) &&
2600 (g_pBTBlocks
[j
- FIRST_BT_ID
] != BTBLOCK_INVAL
))
2602 if (j
== FIRST_BT_ID
) {
2604 last_erased
= LAST_BT_ID
;
2606 last_erased
= (u8
)j
+ 1;
2607 while ((j
> FIRST_BT_ID
) && (BTBLOCK_INVAL
==
2608 g_pBTBlocks
[j
- FIRST_BT_ID
]))
2613 while (g_pBTBlocks
[j
- FIRST_BT_ID
] == BTBLOCK_INVAL
)
2615 last_erased
= (u8
)j
;
2616 while ((j
< LAST_BT_ID
) && (BTBLOCK_INVAL
!=
2617 g_pBTBlocks
[j
- FIRST_BT_ID
]))
2619 if (g_pBTBlocks
[j
-FIRST_BT_ID
] == BTBLOCK_INVAL
)
2623 if (last_erased
> j
)
2624 j
+= (1 + LAST_BT_ID
- FIRST_BT_ID
);
2626 for (; (j
>= last_erased
) && (FAIL
== wResult
); j
--) {
2627 i
= (j
- FIRST_BT_ID
) %
2628 (1 + LAST_BT_ID
- FIRST_BT_ID
);
2630 FTL_Search_Block_Table_IN_Block(g_pBTBlocks
[i
],
2631 i
+ FIRST_BT_ID
, &Page
);
2632 if (g_cBlockTableStatus
== IN_PROGRESS_BLOCK_TABLE
)
2633 block_table_found
= 0;
2635 while ((search_result
== PASS
) && (FAIL
== wResult
)) {
2636 nand_dbg_print(NAND_DBG_DEBUG
,
2637 "FTL_Read_Block_Table:"
2638 "Block: %u Page: %u "
2639 "contains block table\n",
2640 (unsigned int)g_pBTBlocks
[i
],
2641 (unsigned int)Page
);
2643 tempBuf
= tmp_buf2_read_blk_table
;
2645 for (k
= 0; k
< bt_pages
; k
++) {
2646 Block
= g_pBTBlocks
[i
];
2650 GLOB_LLD_Read_Page_Main_Polling(
2651 tempBuf
, Block
, Page
, PageCount
);
2653 tempvar
= k
? 0 : 4;
2656 FTL_Copy_Block_Table_From_Flash(
2658 DeviceInfo
.wPageDataSize
- tempvar
,
2664 wResult
= FTL_Check_Block_Table(FAIL
);
2665 if (FAIL
== wResult
) {
2666 block_table_found
= 0;
2667 if (Page
> bt_pages
)
2668 Page
-= ((bt_pages
<<1) + 1);
2670 search_result
= FAIL
;
2676 if (PASS
== wResult
) {
2677 if (!block_table_found
)
2678 FTL_Execute_SPL_Recovery();
2680 if (g_cBlockTableStatus
== IN_PROGRESS_BLOCK_TABLE
)
2681 g_wBlockTableOffset
= (u16
)Page
+ 1;
2683 g_wBlockTableOffset
= (u16
)Page
- bt_pages
;
2685 g_wBlockTableIndex
= (u32
)g_pBTBlocks
[i
];
2688 if (DeviceInfo
.MLCDevice
)
2689 memcpy(g_pBTStartingCopy
, g_pBlockTable
,
2690 DeviceInfo
.wDataBlockNum
* sizeof(u32
)
2691 + DeviceInfo
.wDataBlockNum
* sizeof(u8
)
2692 + DeviceInfo
.wDataBlockNum
* sizeof(u16
));
2694 memcpy(g_pBTStartingCopy
, g_pBlockTable
,
2695 DeviceInfo
.wDataBlockNum
* sizeof(u32
)
2696 + DeviceInfo
.wDataBlockNum
* sizeof(u8
));
2700 if (FAIL
== wResult
)
2701 printk(KERN_ERR
"Yunpeng - "
2702 "Can not find valid spectra block table!\n");
2704 #if AUTO_FORMAT_FLASH
2705 if (FAIL
== wResult
) {
2706 nand_dbg_print(NAND_DBG_DEBUG
, "doing auto-format\n");
2707 wResult
= FTL_Format_Flash(0);
2714 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2715 * Function: FTL_Get_Page_Num
2716 * Inputs: Size in bytes
2717 * Outputs: Size in pages
2718 * Description: It calculates the pages required for the length passed
2719 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2720 static u32
FTL_Get_Page_Num(u64 length
)
2722 return (u32
)((length
>> DeviceInfo
.nBitsInPageDataSize
) +
2723 (GLOB_u64_Remainder(length
, 1) > 0 ? 1 : 0));
2726 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2727 * Function: FTL_Get_Physical_Block_Addr
2728 * Inputs: Block Address (byte format)
2729 * Outputs: Physical address of the block.
2730 * Description: It translates LBA to PBA by returning address stored
2731 * at the LBA location in the block table
2732 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2733 static u64
FTL_Get_Physical_Block_Addr(u64 logical_addr
)
2738 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2739 __FILE__
, __LINE__
, __func__
);
2741 pbt
= (u32
*)g_pBlockTable
;
2742 physical_addr
= (u64
) DeviceInfo
.wBlockDataSize
*
2743 (pbt
[BLK_FROM_ADDR(logical_addr
)] & (~BAD_BLOCK
));
2745 return physical_addr
;
2748 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2749 * Function: FTL_Get_Block_Index
2750 * Inputs: Physical Block no.
2751 * Outputs: Logical block no. /BAD_BLOCK
2752 * Description: It returns the logical block no. for the PBA passed
2753 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2754 static u32
FTL_Get_Block_Index(u32 wBlockNum
)
2756 u32
*pbt
= (u32
*)g_pBlockTable
;
2759 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
2760 __FILE__
, __LINE__
, __func__
);
2762 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++)
2763 if (wBlockNum
== (pbt
[i
] & (~BAD_BLOCK
)))
2769 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2770 * Function: GLOB_FTL_Wear_Leveling
2773 * Description: This is static wear leveling (done by explicit call)
2774 * do complete static wear leveling
2775 * do complete garbage collection
2776 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2777 int GLOB_FTL_Wear_Leveling(void)
2779 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
2780 __FILE__
, __LINE__
, __func__
);
2782 FTL_Static_Wear_Leveling();
2783 GLOB_FTL_Garbage_Collection();
2788 static void find_least_most_worn(u8
*chg
,
2789 u32
*least_idx
, u8
*least_cnt
,
2790 u32
*most_idx
, u8
*most_cnt
)
2792 u32
*pbt
= (u32
*)g_pBlockTable
;
2797 for (i
= BLOCK_TABLE_INDEX
+ 1; i
< DeviceInfo
.wDataBlockNum
; i
++) {
2798 if (IS_BAD_BLOCK(i
) || PASS
== chg
[i
])
2801 idx
= (u32
) ((~BAD_BLOCK
) & pbt
[i
]);
2802 cnt
= g_pWearCounter
[idx
- DeviceInfo
.wSpectraStartBlock
];
2804 if (IS_SPARE_BLOCK(i
)) {
2805 if (cnt
> *most_cnt
) {
2811 if (IS_DATA_BLOCK(i
)) {
2812 if (cnt
< *least_cnt
) {
2818 if (PASS
== chg
[*most_idx
] || PASS
== chg
[*least_idx
]) {
2819 debug_boundary_error(*most_idx
,
2820 DeviceInfo
.wDataBlockNum
, 0);
2821 debug_boundary_error(*least_idx
,
2822 DeviceInfo
.wDataBlockNum
, 0);
2828 static int move_blks_for_wear_leveling(u8
*chg
,
2829 u32
*least_idx
, u32
*rep_blk_num
, int *result
)
2831 u32
*pbt
= (u32
*)g_pBlockTable
;
2833 int j
, ret_cp_blk
, ret_erase
;
2836 chg
[*least_idx
] = PASS
;
2837 debug_boundary_error(*least_idx
, DeviceInfo
.wDataBlockNum
, 0);
2839 rep_blk
= FTL_Replace_MWBlock();
2840 if (rep_blk
!= BAD_BLOCK
) {
2841 nand_dbg_print(NAND_DBG_DEBUG
,
2842 "More than two spare blocks exist so do it\n");
2843 nand_dbg_print(NAND_DBG_DEBUG
, "Block Replaced is %d\n",
2846 chg
[rep_blk
] = PASS
;
2848 if (IN_PROGRESS_BLOCK_TABLE
!= g_cBlockTableStatus
) {
2849 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
2850 FTL_Write_IN_Progress_Block_Table_Page();
2853 for (j
= 0; j
< RETRY_TIMES
; j
++) {
2854 ret_cp_blk
= FTL_Copy_Block((u64
)(*least_idx
) *
2855 DeviceInfo
.wBlockDataSize
,
2856 (u64
)rep_blk
* DeviceInfo
.wBlockDataSize
);
2857 if (FAIL
== ret_cp_blk
) {
2858 ret_erase
= GLOB_FTL_Block_Erase((u64
)rep_blk
2859 * DeviceInfo
.wBlockDataSize
);
2860 if (FAIL
== ret_erase
)
2861 MARK_BLOCK_AS_BAD(pbt
[rep_blk
]);
2863 nand_dbg_print(NAND_DBG_DEBUG
,
2864 "FTL_Copy_Block == OK\n");
2869 if (j
< RETRY_TIMES
) {
2871 u32 old_idx
= FTL_Get_Block_Index(*least_idx
);
2872 u32 rep_idx
= FTL_Get_Block_Index(rep_blk
);
2873 tmp
= (u32
)(DISCARD_BLOCK
| pbt
[old_idx
]);
2874 pbt
[old_idx
] = (u32
)((~SPARE_BLOCK
) &
2878 p_BTableChangesDelta
= (struct BTableChangesDelta
*)
2880 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
2881 p_BTableChangesDelta
->ftl_cmd_cnt
=
2883 p_BTableChangesDelta
->BT_Index
= old_idx
;
2884 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[old_idx
];
2885 p_BTableChangesDelta
->ValidFields
= 0x0C;
2887 p_BTableChangesDelta
= (struct BTableChangesDelta
*)
2889 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
2891 p_BTableChangesDelta
->ftl_cmd_cnt
=
2893 p_BTableChangesDelta
->BT_Index
= rep_idx
;
2894 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[rep_idx
];
2895 p_BTableChangesDelta
->ValidFields
= 0x0C;
2898 pbt
[FTL_Get_Block_Index(rep_blk
)] |= BAD_BLOCK
;
2900 p_BTableChangesDelta
= (struct BTableChangesDelta
*)
2902 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
2904 p_BTableChangesDelta
->ftl_cmd_cnt
=
2906 p_BTableChangesDelta
->BT_Index
=
2907 FTL_Get_Block_Index(rep_blk
);
2908 p_BTableChangesDelta
->BT_Entry_Value
=
2909 pbt
[FTL_Get_Block_Index(rep_blk
)];
2910 p_BTableChangesDelta
->ValidFields
= 0x0C;
2916 if (((*rep_blk_num
)++) > WEAR_LEVELING_BLOCK_NUM
)
2919 printk(KERN_ERR
"Less than 3 spare blocks exist so quit\n");
2926 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
2927 * Function: FTL_Static_Wear_Leveling
2929 * Outputs: PASS=0 / FAIL=1
2930 * Description: This is static wear leveling (done by explicit call)
2931 * search for most&least used
2932 * if difference < GATE:
2933 * update the block table with exhange
2934 * mark block table in flash as IN_PROGRESS
2936 * the caller should handle GC clean up after calling this function
2937 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
2938 int FTL_Static_Wear_Leveling(void)
2946 u32 replaced_blks
= 0;
2947 u8
*chang_flag
= flags_static_wear_leveling
;
2949 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
2950 __FILE__
, __LINE__
, __func__
);
2955 memset(chang_flag
, FAIL
, DeviceInfo
.wDataBlockNum
);
2956 while (go_on
== PASS
) {
2957 nand_dbg_print(NAND_DBG_DEBUG
,
2958 "starting static wear leveling\n");
2960 least_worn_cnt
= 0xFF;
2961 least_worn_idx
= BLOCK_TABLE_INDEX
;
2962 most_worn_idx
= BLOCK_TABLE_INDEX
;
2964 find_least_most_worn(chang_flag
, &least_worn_idx
,
2965 &least_worn_cnt
, &most_worn_idx
, &most_worn_cnt
);
2967 nand_dbg_print(NAND_DBG_DEBUG
,
2968 "Used and least worn is block %u, whos count is %u\n",
2969 (unsigned int)least_worn_idx
,
2970 (unsigned int)least_worn_cnt
);
2972 nand_dbg_print(NAND_DBG_DEBUG
,
2973 "Free and most worn is block %u, whos count is %u\n",
2974 (unsigned int)most_worn_idx
,
2975 (unsigned int)most_worn_cnt
);
2977 if ((most_worn_cnt
> least_worn_cnt
) &&
2978 (most_worn_cnt
- least_worn_cnt
> WEAR_LEVELING_GATE
))
2979 go_on
= move_blks_for_wear_leveling(chang_flag
,
2980 &least_worn_idx
, &replaced_blks
, &result
);
2989 static int do_garbage_collection(u32 discard_cnt
)
2991 u32
*pbt
= (u32
*)g_pBlockTable
;
2993 u8 bt_block_erased
= 0;
2994 int i
, cnt
, ret
= FAIL
;
2998 while ((i
< DeviceInfo
.wDataBlockNum
) && (discard_cnt
> 0) &&
2999 ((ftl_cmd_cnt
+ 28) < 256)) {
3000 if (((pbt
[i
] & BAD_BLOCK
) != BAD_BLOCK
) &&
3001 (pbt
[i
] & DISCARD_BLOCK
)) {
3002 if (IN_PROGRESS_BLOCK_TABLE
!= g_cBlockTableStatus
) {
3003 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
3004 FTL_Write_IN_Progress_Block_Table_Page();
3007 addr
= FTL_Get_Physical_Block_Addr((u64
)i
*
3008 DeviceInfo
.wBlockDataSize
);
3009 pba
= BLK_FROM_ADDR(addr
);
3011 for (cnt
= FIRST_BT_ID
; cnt
<= LAST_BT_ID
; cnt
++) {
3012 if (pba
== g_pBTBlocks
[cnt
- FIRST_BT_ID
]) {
3013 nand_dbg_print(NAND_DBG_DEBUG
,
3014 "GC will erase BT block %u\n",
3018 bt_block_erased
= 1;
3023 if (bt_block_erased
) {
3024 bt_block_erased
= 0;
3028 addr
= FTL_Get_Physical_Block_Addr((u64
)i
*
3029 DeviceInfo
.wBlockDataSize
);
3031 if (PASS
== GLOB_FTL_Block_Erase(addr
)) {
3032 pbt
[i
] &= (u32
)(~DISCARD_BLOCK
);
3033 pbt
[i
] |= (u32
)(SPARE_BLOCK
);
3034 p_BTableChangesDelta
=
3035 (struct BTableChangesDelta
*)
3038 sizeof(struct BTableChangesDelta
);
3039 p_BTableChangesDelta
->ftl_cmd_cnt
=
3041 p_BTableChangesDelta
->BT_Index
= i
;
3042 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[i
];
3043 p_BTableChangesDelta
->ValidFields
= 0x0C;
3047 MARK_BLOCK_AS_BAD(pbt
[i
]);
3058 static int do_garbage_collection(u32 discard_cnt
)
3060 u32
*pbt
= (u32
*)g_pBlockTable
;
3062 u8 bt_block_erased
= 0;
3063 int i
, cnt
, ret
= FAIL
;
3067 while ((i
< DeviceInfo
.wDataBlockNum
) && (discard_cnt
> 0)) {
3068 if (((pbt
[i
] & BAD_BLOCK
) != BAD_BLOCK
) &&
3069 (pbt
[i
] & DISCARD_BLOCK
)) {
3070 if (IN_PROGRESS_BLOCK_TABLE
!= g_cBlockTableStatus
) {
3071 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
3072 FTL_Write_IN_Progress_Block_Table_Page();
3075 addr
= FTL_Get_Physical_Block_Addr((u64
)i
*
3076 DeviceInfo
.wBlockDataSize
);
3077 pba
= BLK_FROM_ADDR(addr
);
3079 for (cnt
= FIRST_BT_ID
; cnt
<= LAST_BT_ID
; cnt
++) {
3080 if (pba
== g_pBTBlocks
[cnt
- FIRST_BT_ID
]) {
3081 nand_dbg_print(NAND_DBG_DEBUG
,
3082 "GC will erase BT block %d\n",
3086 bt_block_erased
= 1;
3091 if (bt_block_erased
) {
3092 bt_block_erased
= 0;
3096 /* If the discard block is L2 cache block, then just skip it */
3097 for (cnt
= 0; cnt
< BLK_NUM_FOR_L2_CACHE
; cnt
++) {
3098 if (cache_l2
.blk_array
[cnt
] == pba
) {
3099 nand_dbg_print(NAND_DBG_DEBUG
,
3100 "GC will erase L2 cache blk %d\n",
3105 if (cnt
< BLK_NUM_FOR_L2_CACHE
) { /* Skip it */
3111 addr
= FTL_Get_Physical_Block_Addr((u64
)i
*
3112 DeviceInfo
.wBlockDataSize
);
3114 if (PASS
== GLOB_FTL_Block_Erase(addr
)) {
3115 pbt
[i
] &= (u32
)(~DISCARD_BLOCK
);
3116 pbt
[i
] |= (u32
)(SPARE_BLOCK
);
3120 MARK_BLOCK_AS_BAD(pbt
[i
]);
3131 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3132 * Function: GLOB_FTL_Garbage_Collection
3134 * Outputs: PASS / FAIL (returns the number of un-erased blocks
3135 * Description: search the block table for all discarded blocks to erase
3136 * for each discarded block:
3137 * set the flash block to IN_PROGRESS
3139 * update the block table
3140 * write the block table to flash
3141 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3142 int GLOB_FTL_Garbage_Collection(void)
3147 u32
*pbt
= (u32
*)g_pBlockTable
;
3149 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
3150 __FILE__
, __LINE__
, __func__
);
3153 printk(KERN_ALERT
"GLOB_FTL_Garbage_Collection() "
3154 "has been re-entered! Exit.\n");
3160 GLOB_FTL_BT_Garbage_Collection();
3162 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
3163 if (IS_DISCARDED_BLOCK(i
))
3167 if (wDiscard
<= 0) {
3172 nand_dbg_print(NAND_DBG_DEBUG
,
3173 "Found %d discarded blocks\n", wDiscard
);
3175 FTL_Write_Block_Table(FAIL
);
3177 wResult
= do_garbage_collection(wDiscard
);
3179 FTL_Write_Block_Table(FAIL
);
3188 static int do_bt_garbage_collection(void)
3191 u32
*pbt
= (u32
*)g_pBlockTable
;
3192 u32
*pBTBlocksNode
= (u32
*)g_pBTBlocks
;
3196 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3197 __FILE__
, __LINE__
, __func__
);
3204 for (i
= last_erased
; (i
<= LAST_BT_ID
) &&
3205 (g_pBTBlocks
[((i
+ 2) % (1 + LAST_BT_ID
- FIRST_BT_ID
)) +
3206 FIRST_BT_ID
- FIRST_BT_ID
] != BTBLOCK_INVAL
) &&
3207 ((ftl_cmd_cnt
+ 28)) < 256; i
++) {
3208 pba
= pBTBlocksNode
[i
- FIRST_BT_ID
];
3209 lba
= FTL_Get_Block_Index(pba
);
3210 nand_dbg_print(NAND_DBG_DEBUG
,
3211 "do_bt_garbage_collection: pba %d, lba %d\n",
3213 nand_dbg_print(NAND_DBG_DEBUG
,
3214 "Block Table Entry: %d", pbt
[lba
]);
3216 if (((pbt
[lba
] & BAD_BLOCK
) != BAD_BLOCK
) &&
3217 (pbt
[lba
] & DISCARD_BLOCK
)) {
3218 nand_dbg_print(NAND_DBG_DEBUG
,
3219 "do_bt_garbage_collection_cdma: "
3220 "Erasing Block tables present in block %d\n",
3222 addr
= FTL_Get_Physical_Block_Addr((u64
)lba
*
3223 DeviceInfo
.wBlockDataSize
);
3224 if (PASS
== GLOB_FTL_Block_Erase(addr
)) {
3225 pbt
[lba
] &= (u32
)(~DISCARD_BLOCK
);
3226 pbt
[lba
] |= (u32
)(SPARE_BLOCK
);
3228 p_BTableChangesDelta
=
3229 (struct BTableChangesDelta
*)
3232 sizeof(struct BTableChangesDelta
);
3234 p_BTableChangesDelta
->ftl_cmd_cnt
=
3236 p_BTableChangesDelta
->BT_Index
= lba
;
3237 p_BTableChangesDelta
->BT_Entry_Value
=
3240 p_BTableChangesDelta
->ValidFields
= 0x0C;
3243 pBTBlocksNode
[last_erased
- FIRST_BT_ID
] =
3245 nand_dbg_print(NAND_DBG_DEBUG
,
3246 "resetting bt entry at index %d "
3248 pBTBlocksNode
[i
- FIRST_BT_ID
]);
3249 if (last_erased
== LAST_BT_ID
)
3250 last_erased
= FIRST_BT_ID
;
3254 MARK_BLOCK_AS_BAD(pbt
[lba
]);
3265 static int do_bt_garbage_collection(void)
3268 u32
*pbt
= (u32
*)g_pBlockTable
;
3269 u32
*pBTBlocksNode
= (u32
*)g_pBTBlocks
;
3273 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3274 __FILE__
, __LINE__
, __func__
);
3281 for (i
= last_erased
; (i
<= LAST_BT_ID
) &&
3282 (g_pBTBlocks
[((i
+ 2) % (1 + LAST_BT_ID
- FIRST_BT_ID
)) +
3283 FIRST_BT_ID
- FIRST_BT_ID
] != BTBLOCK_INVAL
); i
++) {
3284 pba
= pBTBlocksNode
[i
- FIRST_BT_ID
];
3285 lba
= FTL_Get_Block_Index(pba
);
3286 nand_dbg_print(NAND_DBG_DEBUG
,
3287 "do_bt_garbage_collection_cdma: pba %d, lba %d\n",
3289 nand_dbg_print(NAND_DBG_DEBUG
,
3290 "Block Table Entry: %d", pbt
[lba
]);
3292 if (((pbt
[lba
] & BAD_BLOCK
) != BAD_BLOCK
) &&
3293 (pbt
[lba
] & DISCARD_BLOCK
)) {
3294 nand_dbg_print(NAND_DBG_DEBUG
,
3295 "do_bt_garbage_collection: "
3296 "Erasing Block tables present in block %d\n",
3298 addr
= FTL_Get_Physical_Block_Addr((u64
)lba
*
3299 DeviceInfo
.wBlockDataSize
);
3300 if (PASS
== GLOB_FTL_Block_Erase(addr
)) {
3301 pbt
[lba
] &= (u32
)(~DISCARD_BLOCK
);
3302 pbt
[lba
] |= (u32
)(SPARE_BLOCK
);
3304 pBTBlocksNode
[last_erased
- FIRST_BT_ID
] =
3306 nand_dbg_print(NAND_DBG_DEBUG
,
3307 "resetting bt entry at index %d "
3309 pBTBlocksNode
[i
- FIRST_BT_ID
]);
3310 if (last_erased
== LAST_BT_ID
)
3311 last_erased
= FIRST_BT_ID
;
3315 MARK_BLOCK_AS_BAD(pbt
[lba
]);
3327 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3328 * Function: GLOB_FTL_BT_Garbage_Collection
3330 * Outputs: PASS / FAIL (returns the number of un-erased blocks
3331 * Description: Erases discarded blocks containing Block table
3333 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3334 int GLOB_FTL_BT_Garbage_Collection(void)
3336 return do_bt_garbage_collection();
3339 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3340 * Function: FTL_Replace_OneBlock
3341 * Inputs: Block number 1
3343 * Outputs: Replaced Block Number
3344 * Description: Interchange block table entries at wBlockNum and wReplaceNum
3346 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3347 static u32
FTL_Replace_OneBlock(u32 blk
, u32 rep_blk
)
3350 u32 replace_node
= BAD_BLOCK
;
3351 u32
*pbt
= (u32
*)g_pBlockTable
;
3353 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3354 __FILE__
, __LINE__
, __func__
);
3356 if (rep_blk
!= BAD_BLOCK
) {
3357 if (IS_BAD_BLOCK(blk
))
3360 tmp_blk
= DISCARD_BLOCK
| (~SPARE_BLOCK
& pbt
[blk
]);
3362 replace_node
= (u32
) ((~SPARE_BLOCK
) & pbt
[rep_blk
]);
3363 pbt
[blk
] = replace_node
;
3364 pbt
[rep_blk
] = tmp_blk
;
3367 p_BTableChangesDelta
=
3368 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3369 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3371 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
3372 p_BTableChangesDelta
->BT_Index
= blk
;
3373 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[blk
];
3375 p_BTableChangesDelta
->ValidFields
= 0x0C;
3377 p_BTableChangesDelta
=
3378 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3379 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3381 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
3382 p_BTableChangesDelta
->BT_Index
= rep_blk
;
3383 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[rep_blk
];
3384 p_BTableChangesDelta
->ValidFields
= 0x0C;
3388 return replace_node
;
3391 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3392 * Function: FTL_Write_Block_Table_Data
3393 * Inputs: Block table size in pages
3394 * Outputs: PASS=0 / FAIL=1
3395 * Description: Write block table data in flash
3396 * If first page and last page
3397 * Write data+BT flag
3400 * BT flag is a counter. Its value is incremented for block table
3401 * write in a new Block
3402 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3403 static int FTL_Write_Block_Table_Data(void)
3405 u64 dwBlockTableAddr
, pTempAddr
;
3407 u16 Page
, PageCount
;
3408 u8
*tempBuf
= tmp_buf_write_blk_table_data
;
3412 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3413 __FILE__
, __LINE__
, __func__
);
3416 (u64
)((u64
)g_wBlockTableIndex
* DeviceInfo
.wBlockDataSize
+
3417 (u64
)g_wBlockTableOffset
* DeviceInfo
.wPageDataSize
);
3418 pTempAddr
= dwBlockTableAddr
;
3420 bt_pages
= FTL_Get_Block_Table_Flash_Size_Pages();
3422 nand_dbg_print(NAND_DBG_DEBUG
, "FTL_Write_Block_Table_Data: "
3423 "page= %d BlockTableIndex= %d "
3424 "BlockTableOffset=%d\n", bt_pages
,
3425 g_wBlockTableIndex
, g_wBlockTableOffset
);
3427 Block
= BLK_FROM_ADDR(pTempAddr
);
3428 Page
= PAGE_FROM_ADDR(pTempAddr
, Block
);
3431 if (bt_block_changed
) {
3432 if (bt_flag
== LAST_BT_ID
) {
3433 bt_flag
= FIRST_BT_ID
;
3434 g_pBTBlocks
[bt_flag
- FIRST_BT_ID
] = Block
;
3435 } else if (bt_flag
< LAST_BT_ID
) {
3437 g_pBTBlocks
[bt_flag
- FIRST_BT_ID
] = Block
;
3440 if ((bt_flag
> (LAST_BT_ID
-4)) &&
3441 g_pBTBlocks
[FIRST_BT_ID
- FIRST_BT_ID
] !=
3443 bt_block_changed
= 0;
3444 GLOB_FTL_BT_Garbage_Collection();
3447 bt_block_changed
= 0;
3448 nand_dbg_print(NAND_DBG_DEBUG
,
3449 "Block Table Counter is %u Block %u\n",
3450 bt_flag
, (unsigned int)Block
);
3453 memset(tempBuf
, 0, 3);
3454 tempBuf
[3] = bt_flag
;
3455 wBytesCopied
= FTL_Copy_Block_Table_To_Flash(tempBuf
+ 4,
3456 DeviceInfo
.wPageDataSize
- 4, 0);
3457 memset(&tempBuf
[wBytesCopied
+ 4], 0xff,
3458 DeviceInfo
.wPageSize
- (wBytesCopied
+ 4));
3459 FTL_Insert_Block_Table_Signature(&tempBuf
[DeviceInfo
.wPageDataSize
],
3463 memcpy(g_pNextBlockTable
, tempBuf
,
3464 DeviceInfo
.wPageSize
* sizeof(u8
));
3465 nand_dbg_print(NAND_DBG_DEBUG
, "Writing First Page of Block Table "
3466 "Block %u Page %u\n", (unsigned int)Block
, Page
);
3467 if (FAIL
== GLOB_LLD_Write_Page_Main_Spare_cdma(g_pNextBlockTable
,
3469 LLD_CMD_FLAG_MODE_CDMA
| LLD_CMD_FLAG_ORDER_BEFORE_REST
)) {
3470 nand_dbg_print(NAND_DBG_WARN
, "NAND Program fail in "
3471 "%s, Line %d, Function: %s, "
3472 "new Bad Block %d generated!\n",
3473 __FILE__
, __LINE__
, __func__
, Block
);
3478 g_pNextBlockTable
+= ((DeviceInfo
.wPageSize
* sizeof(u8
)));
3480 if (FAIL
== GLOB_LLD_Write_Page_Main_Spare(tempBuf
, Block
, Page
, 1)) {
3481 nand_dbg_print(NAND_DBG_WARN
,
3482 "NAND Program fail in %s, Line %d, Function: %s, "
3483 "new Bad Block %d generated!\n",
3484 __FILE__
, __LINE__
, __func__
, Block
);
3490 PageCount
= bt_pages
- 1;
3491 if (PageCount
> 1) {
3492 wBytesCopied
+= FTL_Copy_Block_Table_To_Flash(tempBuf
,
3493 DeviceInfo
.wPageDataSize
* (PageCount
- 1),
3497 memcpy(g_pNextBlockTable
, tempBuf
,
3498 (PageCount
- 1) * DeviceInfo
.wPageDataSize
);
3499 if (FAIL
== GLOB_LLD_Write_Page_Main_cdma(
3500 g_pNextBlockTable
, Block
, Page
+ 1,
3502 nand_dbg_print(NAND_DBG_WARN
,
3503 "NAND Program fail in %s, Line %d, "
3505 "new Bad Block %d generated!\n",
3506 __FILE__
, __LINE__
, __func__
,
3512 g_pNextBlockTable
+= (PageCount
- 1) *
3513 DeviceInfo
.wPageDataSize
* sizeof(u8
);
3515 if (FAIL
== GLOB_LLD_Write_Page_Main(tempBuf
,
3516 Block
, Page
+ 1, PageCount
- 1)) {
3517 nand_dbg_print(NAND_DBG_WARN
,
3518 "NAND Program fail in %s, Line %d, "
3520 "new Bad Block %d generated!\n",
3521 __FILE__
, __LINE__
, __func__
,
3528 wBytesCopied
= FTL_Copy_Block_Table_To_Flash(tempBuf
,
3529 DeviceInfo
.wPageDataSize
, wBytesCopied
);
3530 memset(&tempBuf
[wBytesCopied
], 0xff,
3531 DeviceInfo
.wPageSize
-wBytesCopied
);
3532 FTL_Insert_Block_Table_Signature(
3533 &tempBuf
[DeviceInfo
.wPageDataSize
], bt_flag
);
3535 memcpy(g_pNextBlockTable
, tempBuf
,
3536 DeviceInfo
.wPageSize
* sizeof(u8
));
3537 nand_dbg_print(NAND_DBG_DEBUG
,
3538 "Writing the last Page of Block Table "
3539 "Block %u Page %u\n",
3540 (unsigned int)Block
, Page
+ bt_pages
- 1);
3541 if (FAIL
== GLOB_LLD_Write_Page_Main_Spare_cdma(
3542 g_pNextBlockTable
, Block
, Page
+ bt_pages
- 1, 1,
3543 LLD_CMD_FLAG_MODE_CDMA
|
3544 LLD_CMD_FLAG_ORDER_BEFORE_REST
)) {
3545 nand_dbg_print(NAND_DBG_WARN
,
3546 "NAND Program fail in %s, Line %d, "
3547 "Function: %s, new Bad Block %d generated!\n",
3548 __FILE__
, __LINE__
, __func__
, Block
);
3553 if (FAIL
== GLOB_LLD_Write_Page_Main_Spare(tempBuf
,
3554 Block
, Page
+bt_pages
- 1, 1)) {
3555 nand_dbg_print(NAND_DBG_WARN
,
3556 "NAND Program fail in %s, Line %d, "
3558 "new Bad Block %d generated!\n",
3559 __FILE__
, __LINE__
, __func__
, Block
);
3565 nand_dbg_print(NAND_DBG_DEBUG
, "FTL_Write_Block_Table_Data: done\n");
3571 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3572 * Function: FTL_Replace_Block_Table
3574 * Outputs: PASS=0 / FAIL=1
3575 * Description: Get a new block to write block table
3576 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3577 static u32
FTL_Replace_Block_Table(void)
3582 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3583 __FILE__
, __LINE__
, __func__
);
3585 blk
= FTL_Replace_LWBlock(BLOCK_TABLE_INDEX
, &gc
);
3587 if ((BAD_BLOCK
== blk
) && (PASS
== gc
)) {
3588 GLOB_FTL_Garbage_Collection();
3589 blk
= FTL_Replace_LWBlock(BLOCK_TABLE_INDEX
, &gc
);
3591 if (BAD_BLOCK
== blk
)
3592 printk(KERN_ERR
"%s, %s: There is no spare block. "
3593 "It should never happen\n",
3594 __FILE__
, __func__
);
3596 nand_dbg_print(NAND_DBG_DEBUG
, "New Block table Block is %d\n", blk
);
3601 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3602 * Function: FTL_Replace_LWBlock
3603 * Inputs: Block number
3604 * Pointer to Garbage Collect flag
3606 * Description: Determine the least weared block by traversing
3608 * Set Garbage collection to be called if number of spare
3609 * block is less than Free Block Gate count
3610 * Change Block table entry to map least worn block for current
3612 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3613 static u32
FTL_Replace_LWBlock(u32 wBlockNum
, int *pGarbageCollect
)
3616 u32
*pbt
= (u32
*)g_pBlockTable
;
3617 u8 wLeastWornCounter
= 0xFF;
3618 u32 wLeastWornIndex
= BAD_BLOCK
;
3619 u32 wSpareBlockNum
= 0;
3620 u32 wDiscardBlockNum
= 0;
3622 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3623 __FILE__
, __LINE__
, __func__
);
3625 if (IS_SPARE_BLOCK(wBlockNum
)) {
3626 *pGarbageCollect
= FAIL
;
3627 pbt
[wBlockNum
] = (u32
)(pbt
[wBlockNum
] & (~SPARE_BLOCK
));
3629 p_BTableChangesDelta
=
3630 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3631 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3632 p_BTableChangesDelta
->ftl_cmd_cnt
=
3634 p_BTableChangesDelta
->BT_Index
= (u32
)(wBlockNum
);
3635 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[wBlockNum
];
3636 p_BTableChangesDelta
->ValidFields
= 0x0C;
3638 return pbt
[wBlockNum
];
3641 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
3642 if (IS_DISCARDED_BLOCK(i
))
3645 if (IS_SPARE_BLOCK(i
)) {
3646 u32 wPhysicalIndex
= (u32
)((~BAD_BLOCK
) & pbt
[i
]);
3647 if (wPhysicalIndex
> DeviceInfo
.wSpectraEndBlock
)
3648 printk(KERN_ERR
"FTL_Replace_LWBlock: "
3649 "This should never occur!\n");
3650 if (g_pWearCounter
[wPhysicalIndex
-
3651 DeviceInfo
.wSpectraStartBlock
] <
3652 wLeastWornCounter
) {
3654 g_pWearCounter
[wPhysicalIndex
-
3655 DeviceInfo
.wSpectraStartBlock
];
3656 wLeastWornIndex
= i
;
3662 nand_dbg_print(NAND_DBG_WARN
,
3663 "FTL_Replace_LWBlock: Least Worn Counter %d\n",
3664 (int)wLeastWornCounter
);
3666 if ((wDiscardBlockNum
>= NUM_FREE_BLOCKS_GATE
) ||
3667 (wSpareBlockNum
<= NUM_FREE_BLOCKS_GATE
))
3668 *pGarbageCollect
= PASS
;
3670 *pGarbageCollect
= FAIL
;
3672 nand_dbg_print(NAND_DBG_DEBUG
,
3673 "FTL_Replace_LWBlock: Discarded Blocks %u Spare"
3675 (unsigned int)wDiscardBlockNum
,
3676 (unsigned int)wSpareBlockNum
);
3678 return FTL_Replace_OneBlock(wBlockNum
, wLeastWornIndex
);
3681 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3682 * Function: FTL_Replace_MWBlock
3684 * Outputs: most worn spare block no./BAD_BLOCK
3685 * Description: It finds most worn spare block.
3686 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3687 static u32
FTL_Replace_MWBlock(void)
3690 u32
*pbt
= (u32
*)g_pBlockTable
;
3691 u8 wMostWornCounter
= 0;
3692 u32 wMostWornIndex
= BAD_BLOCK
;
3693 u32 wSpareBlockNum
= 0;
3695 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3696 __FILE__
, __LINE__
, __func__
);
3698 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
3699 if (IS_SPARE_BLOCK(i
)) {
3700 u32 wPhysicalIndex
= (u32
)((~SPARE_BLOCK
) & pbt
[i
]);
3701 if (g_pWearCounter
[wPhysicalIndex
-
3702 DeviceInfo
.wSpectraStartBlock
] >
3705 g_pWearCounter
[wPhysicalIndex
-
3706 DeviceInfo
.wSpectraStartBlock
];
3707 wMostWornIndex
= wPhysicalIndex
;
3713 if (wSpareBlockNum
<= 2)
3716 return wMostWornIndex
;
3719 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3720 * Function: FTL_Replace_Block
3721 * Inputs: Block Address
3722 * Outputs: PASS=0 / FAIL=1
3723 * Description: If block specified by blk_addr parameter is not free,
3724 * replace it with the least worn block.
3725 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3726 static int FTL_Replace_Block(u64 blk_addr
)
3728 u32 current_blk
= BLK_FROM_ADDR(blk_addr
);
3729 u32
*pbt
= (u32
*)g_pBlockTable
;
3731 int GarbageCollect
= FAIL
;
3733 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3734 __FILE__
, __LINE__
, __func__
);
3736 if (IS_SPARE_BLOCK(current_blk
)) {
3737 pbt
[current_blk
] = (~SPARE_BLOCK
) & pbt
[current_blk
];
3739 p_BTableChangesDelta
=
3740 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3741 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3742 p_BTableChangesDelta
->ftl_cmd_cnt
=
3744 p_BTableChangesDelta
->BT_Index
= current_blk
;
3745 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[current_blk
];
3746 p_BTableChangesDelta
->ValidFields
= 0x0C ;
3751 FTL_Replace_LWBlock(current_blk
, &GarbageCollect
);
3753 if (PASS
== GarbageCollect
)
3754 wResult
= GLOB_FTL_Garbage_Collection();
3759 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3760 * Function: GLOB_FTL_Is_BadBlock
3761 * Inputs: block number to test
3762 * Outputs: PASS (block is BAD) / FAIL (block is not bad)
3763 * Description: test if this block number is flagged as bad
3764 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3765 int GLOB_FTL_Is_BadBlock(u32 wBlockNum
)
3767 u32
*pbt
= (u32
*)g_pBlockTable
;
3769 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3770 __FILE__
, __LINE__
, __func__
);
3772 if (wBlockNum
>= DeviceInfo
.wSpectraStartBlock
3773 && BAD_BLOCK
== (pbt
[wBlockNum
] & BAD_BLOCK
))
3779 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3780 * Function: GLOB_FTL_Flush_Cache
3782 * Outputs: PASS=0 / FAIL=1
3783 * Description: flush all the cache blocks to flash
3784 * if a cache block is not dirty, don't do anything with it
3785 * else, write the block and update the block table
3786 * Note: This function should be called at shutdown/power down.
3787 * to write important data into device
3788 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3789 int GLOB_FTL_Flush_Cache(void)
3793 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
3794 __FILE__
, __LINE__
, __func__
);
3796 for (i
= 0; i
< CACHE_ITEM_NUM
; i
++) {
3797 if (SET
== Cache
.array
[i
].changed
) {
3799 #if RESTORE_CACHE_ON_CDMA_CHAIN_FAILURE
3800 int_cache
[ftl_cmd_cnt
].item
= i
;
3801 int_cache
[ftl_cmd_cnt
].cache
.address
=
3802 Cache
.array
[i
].address
;
3803 int_cache
[ftl_cmd_cnt
].cache
.changed
= CLEAR
;
3806 ret
= write_back_to_l2_cache(Cache
.array
[i
].buf
, Cache
.array
[i
].address
);
3808 Cache
.array
[i
].changed
= CLEAR
;
3810 printk(KERN_ALERT
"Failed when write back to L2 cache!\n");
3811 /* TODO - How to handle this? */
3818 return FTL_Write_Block_Table(FAIL
);
3821 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3822 * Function: GLOB_FTL_Page_Read
3823 * Inputs: pointer to data
3824 * logical address of data (u64 is LBA * Bytes/Page)
3825 * Outputs: PASS=0 / FAIL=1
3826 * Description: reads a page of data into RAM from the cache
3827 * if the data is not already in cache, read from flash to cache
3828 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3829 int GLOB_FTL_Page_Read(u8
*data
, u64 logical_addr
)
3834 nand_dbg_print(NAND_DBG_DEBUG
, "GLOB_FTL_Page_Read - "
3835 "page_addr: %llu\n", logical_addr
);
3837 cache_item
= FTL_Cache_If_Hit(logical_addr
);
3839 if (UNHIT_CACHE_ITEM
== cache_item
) {
3840 nand_dbg_print(NAND_DBG_DEBUG
,
3841 "GLOB_FTL_Page_Read: Cache not hit\n");
3842 res
= FTL_Cache_Write();
3843 if (ERR
== FTL_Cache_Read(logical_addr
))
3845 cache_item
= Cache
.LRU
;
3848 FTL_Cache_Read_Page(data
, logical_addr
, cache_item
);
3853 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3854 * Function: GLOB_FTL_Page_Write
3855 * Inputs: pointer to data
3856 * address of data (ADDRESSTYPE is LBA * Bytes/Page)
3857 * Outputs: PASS=0 / FAIL=1
3858 * Description: writes a page of data from RAM to the cache
3859 * if the data is not already in cache, write back the
3860 * least recently used block and read the addressed block
3861 * from flash to cache
3862 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3863 int GLOB_FTL_Page_Write(u8
*pData
, u64 dwPageAddr
)
3866 u32
*pbt
= (u32
*)g_pBlockTable
;
3869 nand_dbg_print(NAND_DBG_TRACE
, "GLOB_FTL_Page_Write - "
3870 "dwPageAddr: %llu\n", dwPageAddr
);
3872 cache_blk
= FTL_Cache_If_Hit(dwPageAddr
);
3874 if (UNHIT_CACHE_ITEM
== cache_blk
) {
3875 wResult
= FTL_Cache_Write();
3876 if (IS_BAD_BLOCK(BLK_FROM_ADDR(dwPageAddr
))) {
3877 wResult
= FTL_Replace_Block(dwPageAddr
);
3878 pbt
[BLK_FROM_ADDR(dwPageAddr
)] |= SPARE_BLOCK
;
3879 if (wResult
== FAIL
)
3882 if (ERR
== FTL_Cache_Read(dwPageAddr
))
3884 cache_blk
= Cache
.LRU
;
3885 FTL_Cache_Write_Page(pData
, dwPageAddr
, cache_blk
, 0);
3888 FTL_Cache_Write_Page(pData
, dwPageAddr
, cache_blk
,
3889 LLD_CMD_FLAG_ORDER_BEFORE_REST
);
3891 FTL_Cache_Write_Page(pData
, dwPageAddr
, cache_blk
, 0);
3898 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3899 * Function: GLOB_FTL_Block_Erase
3900 * Inputs: address of block to erase (now in byte format, should change to
3902 * Outputs: PASS=0 / FAIL=1
3903 * Description: erases the specified block
3904 * increments the erase count
3905 * If erase count reaches its upper limit,call function to
3906 * do the adjustment as per the relative erase count values
3907 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3908 int GLOB_FTL_Block_Erase(u64 blk_addr
)
3913 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
3914 __FILE__
, __LINE__
, __func__
);
3916 BlkIdx
= (u32
)(blk_addr
>> DeviceInfo
.nBitsInBlockDataSize
);
3918 if (BlkIdx
< DeviceInfo
.wSpectraStartBlock
) {
3919 printk(KERN_ERR
"GLOB_FTL_Block_Erase: "
3920 "This should never occur\n");
3925 status
= GLOB_LLD_Erase_Block_cdma(BlkIdx
, LLD_CMD_FLAG_MODE_CDMA
);
3927 nand_dbg_print(NAND_DBG_WARN
,
3928 "NAND Program fail in %s, Line %d, "
3929 "Function: %s, new Bad Block %d generated!\n",
3930 __FILE__
, __LINE__
, __func__
, BlkIdx
);
3932 status
= GLOB_LLD_Erase_Block(BlkIdx
);
3933 if (status
== FAIL
) {
3934 nand_dbg_print(NAND_DBG_WARN
,
3935 "NAND Program fail in %s, Line %d, "
3936 "Function: %s, new Bad Block %d generated!\n",
3937 __FILE__
, __LINE__
, __func__
, BlkIdx
);
3942 if (DeviceInfo
.MLCDevice
) {
3943 g_pReadCounter
[BlkIdx
- DeviceInfo
.wSpectraStartBlock
] = 0;
3944 if (g_cBlockTableStatus
!= IN_PROGRESS_BLOCK_TABLE
) {
3945 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
3946 FTL_Write_IN_Progress_Block_Table_Page();
3950 g_pWearCounter
[BlkIdx
- DeviceInfo
.wSpectraStartBlock
]++;
3953 p_BTableChangesDelta
=
3954 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3955 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3956 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
3957 p_BTableChangesDelta
->WC_Index
=
3958 BlkIdx
- DeviceInfo
.wSpectraStartBlock
;
3959 p_BTableChangesDelta
->WC_Entry_Value
=
3960 g_pWearCounter
[BlkIdx
- DeviceInfo
.wSpectraStartBlock
];
3961 p_BTableChangesDelta
->ValidFields
= 0x30;
3963 if (DeviceInfo
.MLCDevice
) {
3964 p_BTableChangesDelta
=
3965 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
3966 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
3967 p_BTableChangesDelta
->ftl_cmd_cnt
=
3969 p_BTableChangesDelta
->RC_Index
=
3970 BlkIdx
- DeviceInfo
.wSpectraStartBlock
;
3971 p_BTableChangesDelta
->RC_Entry_Value
=
3972 g_pReadCounter
[BlkIdx
-
3973 DeviceInfo
.wSpectraStartBlock
];
3974 p_BTableChangesDelta
->ValidFields
= 0xC0;
3980 if (g_pWearCounter
[BlkIdx
- DeviceInfo
.wSpectraStartBlock
] == 0xFE)
3981 FTL_Adjust_Relative_Erase_Count(BlkIdx
);
3987 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
3988 * Function: FTL_Adjust_Relative_Erase_Count
3989 * Inputs: index to block that was just incremented and is at the max
3990 * Outputs: PASS=0 / FAIL=1
3991 * Description: If any erase counts at MAX, adjusts erase count of every
3992 * block by subtracting least worn
3993 * counter from counter value of every entry in wear table
3994 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
3995 static int FTL_Adjust_Relative_Erase_Count(u32 Index_of_MAX
)
3997 u8 wLeastWornCounter
= MAX_BYTE_VALUE
;
4000 u32
*pbt
= (u32
*)g_pBlockTable
;
4003 nand_dbg_print(NAND_DBG_TRACE
, "%s, Line %d, Function: %s\n",
4004 __FILE__
, __LINE__
, __func__
);
4006 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++) {
4007 if (IS_BAD_BLOCK(i
))
4009 wWearIndex
= (u32
)(pbt
[i
] & (~BAD_BLOCK
));
4011 if ((wWearIndex
- DeviceInfo
.wSpectraStartBlock
) < 0)
4012 printk(KERN_ERR
"FTL_Adjust_Relative_Erase_Count:"
4013 "This should never occur\n");
4014 wWearCounter
= g_pWearCounter
[wWearIndex
-
4015 DeviceInfo
.wSpectraStartBlock
];
4016 if (wWearCounter
< wLeastWornCounter
)
4017 wLeastWornCounter
= wWearCounter
;
4020 if (wLeastWornCounter
== 0) {
4021 nand_dbg_print(NAND_DBG_WARN
,
4022 "Adjusting Wear Levelling Counters: Special Case\n");
4023 g_pWearCounter
[Index_of_MAX
-
4024 DeviceInfo
.wSpectraStartBlock
]--;
4026 p_BTableChangesDelta
=
4027 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
4028 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
4029 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
4030 p_BTableChangesDelta
->WC_Index
=
4031 Index_of_MAX
- DeviceInfo
.wSpectraStartBlock
;
4032 p_BTableChangesDelta
->WC_Entry_Value
=
4033 g_pWearCounter
[Index_of_MAX
-
4034 DeviceInfo
.wSpectraStartBlock
];
4035 p_BTableChangesDelta
->ValidFields
= 0x30;
4037 FTL_Static_Wear_Leveling();
4039 for (i
= 0; i
< DeviceInfo
.wDataBlockNum
; i
++)
4040 if (!IS_BAD_BLOCK(i
)) {
4041 wWearIndex
= (u32
)(pbt
[i
] & (~BAD_BLOCK
));
4042 g_pWearCounter
[wWearIndex
-
4043 DeviceInfo
.wSpectraStartBlock
] =
4046 DeviceInfo
.wSpectraStartBlock
] -
4049 p_BTableChangesDelta
=
4050 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
4052 sizeof(struct BTableChangesDelta
);
4054 p_BTableChangesDelta
->ftl_cmd_cnt
=
4056 p_BTableChangesDelta
->WC_Index
= wWearIndex
-
4057 DeviceInfo
.wSpectraStartBlock
;
4058 p_BTableChangesDelta
->WC_Entry_Value
=
4059 g_pWearCounter
[wWearIndex
-
4060 DeviceInfo
.wSpectraStartBlock
];
4061 p_BTableChangesDelta
->ValidFields
= 0x30;
4069 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
4070 * Function: FTL_Write_IN_Progress_Block_Table_Page
4073 * Description: It writes in-progress flag page to the page next to
4075 ***********************************************************************/
4076 static int FTL_Write_IN_Progress_Block_Table_Page(void)
4083 u32
*pbt
= (u32
*)g_pBlockTable
;
4084 u32 wTempBlockTableIndex
;
4087 nand_dbg_print(NAND_DBG_WARN
, "%s, Line %d, Function: %s\n",
4088 __FILE__
, __LINE__
, __func__
);
4090 bt_pages
= FTL_Get_Block_Table_Flash_Size_Pages();
4092 dwIPFPageAddr
= g_wBlockTableOffset
+ bt_pages
;
4094 nand_dbg_print(NAND_DBG_DEBUG
, "Writing IPF at "
4095 "Block %d Page %d\n",
4096 g_wBlockTableIndex
, dwIPFPageAddr
);
4099 wResult
= GLOB_LLD_Write_Page_Main_Spare_cdma(g_pIPF
,
4100 g_wBlockTableIndex
, dwIPFPageAddr
, 1,
4101 LLD_CMD_FLAG_MODE_CDMA
| LLD_CMD_FLAG_ORDER_BEFORE_REST
);
4102 if (wResult
== FAIL
) {
4103 nand_dbg_print(NAND_DBG_WARN
,
4104 "NAND Program fail in %s, Line %d, "
4105 "Function: %s, new Bad Block %d generated!\n",
4106 __FILE__
, __LINE__
, __func__
,
4107 g_wBlockTableIndex
);
4109 g_wBlockTableOffset
= dwIPFPageAddr
+ 1;
4110 p_BTableChangesDelta
= (struct BTableChangesDelta
*)g_pBTDelta_Free
;
4111 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
4112 p_BTableChangesDelta
->ftl_cmd_cnt
= ftl_cmd_cnt
;
4113 p_BTableChangesDelta
->g_wBlockTableOffset
= g_wBlockTableOffset
;
4114 p_BTableChangesDelta
->ValidFields
= 0x01;
4117 wResult
= GLOB_LLD_Write_Page_Main_Spare(g_pIPF
,
4118 g_wBlockTableIndex
, dwIPFPageAddr
, 1);
4119 if (wResult
== FAIL
) {
4120 nand_dbg_print(NAND_DBG_WARN
,
4121 "NAND Program fail in %s, Line %d, "
4122 "Function: %s, new Bad Block %d generated!\n",
4123 __FILE__
, __LINE__
, __func__
,
4124 (int)g_wBlockTableIndex
);
4125 MARK_BLOCK_AS_BAD(pbt
[BLOCK_TABLE_INDEX
]);
4126 wTempBlockTableIndex
= FTL_Replace_Block_Table();
4127 bt_block_changed
= 1;
4128 if (BAD_BLOCK
== wTempBlockTableIndex
)
4130 g_wBlockTableIndex
= wTempBlockTableIndex
;
4131 g_wBlockTableOffset
= 0;
4132 /* Block table tag is '00'. Means it's used one */
4133 pbt
[BLOCK_TABLE_INDEX
] = g_wBlockTableIndex
;
4136 g_wBlockTableOffset
= dwIPFPageAddr
+ 1;
4141 /*&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
4142 * Function: FTL_Read_Disturbance
4143 * Inputs: block address
4144 * Outputs: PASS=0 / FAIL=1
4145 * Description: used to handle read disturbance. Data in block that
4146 * reaches its read limit is moved to new block
4147 *&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&*/
4148 int FTL_Read_Disturbance(u32 blk_addr
)
4151 u32
*pbt
= (u32
*) g_pBlockTable
;
4152 u32 dwOldBlockAddr
= blk_addr
;
4155 u32 wLeastReadCounter
= 0xFFFF;
4156 u32 wLeastReadIndex
= BAD_BLOCK
;
4157 u32 wSpareBlockNum
= 0;
4162 nand_dbg_print(NAND_DBG_DEBUG
, "%s, Line %d, Function: %s\n",
4163 __FILE__
, __LINE__
, __func__
);
4166 g_pTempBuf
= cp_back_buf_copies
[cp_back_buf_idx
];
4168 if (cp_back_buf_idx
> COPY_BACK_BUF_NUM
) {
4169 printk(KERN_ERR
"cp_back_buf_copies overflow! Exit."
4170 "Maybe too many pending commands in your CDMA chain.\n");
4174 g_pTempBuf
= tmp_buf_read_disturbance
;
4177 wBlockNum
= FTL_Get_Block_Index(blk_addr
);
4180 /* This is a bug.Here 'i' should be logical block number
4181 * and start from 1 (0 is reserved for block table).
4182 * Have fixed it. - Yunpeng 2008. 12. 19
4184 for (i
= 1; i
< DeviceInfo
.wDataBlockNum
; i
++) {
4185 if (IS_SPARE_BLOCK(i
)) {
4186 u32 wPhysicalIndex
=
4187 (u32
)((~SPARE_BLOCK
) & pbt
[i
]);
4188 if (g_pReadCounter
[wPhysicalIndex
-
4189 DeviceInfo
.wSpectraStartBlock
] <
4190 wLeastReadCounter
) {
4192 g_pReadCounter
[wPhysicalIndex
-
4193 DeviceInfo
.wSpectraStartBlock
];
4194 wLeastReadIndex
= i
;
4200 if (wSpareBlockNum
<= NUM_FREE_BLOCKS_GATE
) {
4201 wResult
= GLOB_FTL_Garbage_Collection();
4202 if (PASS
== wResult
)
4207 wTempNode
= (u32
)(DISCARD_BLOCK
| pbt
[wBlockNum
]);
4208 wReplacedNode
= (u32
)((~SPARE_BLOCK
) &
4209 pbt
[wLeastReadIndex
]);
4211 pbt
[wBlockNum
] = wReplacedNode
;
4212 pbt
[wLeastReadIndex
] = wTempNode
;
4213 p_BTableChangesDelta
=
4214 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
4215 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
4217 p_BTableChangesDelta
->ftl_cmd_cnt
=
4219 p_BTableChangesDelta
->BT_Index
= wBlockNum
;
4220 p_BTableChangesDelta
->BT_Entry_Value
= pbt
[wBlockNum
];
4221 p_BTableChangesDelta
->ValidFields
= 0x0C;
4223 p_BTableChangesDelta
=
4224 (struct BTableChangesDelta
*)g_pBTDelta_Free
;
4225 g_pBTDelta_Free
+= sizeof(struct BTableChangesDelta
);
4227 p_BTableChangesDelta
->ftl_cmd_cnt
=
4229 p_BTableChangesDelta
->BT_Index
= wLeastReadIndex
;
4230 p_BTableChangesDelta
->BT_Entry_Value
=
4231 pbt
[wLeastReadIndex
];
4232 p_BTableChangesDelta
->ValidFields
= 0x0C;
4234 wResult
= GLOB_LLD_Read_Page_Main_cdma(g_pTempBuf
,
4235 dwOldBlockAddr
, 0, DeviceInfo
.wPagesPerBlock
,
4236 LLD_CMD_FLAG_MODE_CDMA
);
4237 if (wResult
== FAIL
)
4242 if (wResult
!= FAIL
) {
4243 if (FAIL
== GLOB_LLD_Write_Page_Main_cdma(
4244 g_pTempBuf
, pbt
[wBlockNum
], 0,
4245 DeviceInfo
.wPagesPerBlock
)) {
4246 nand_dbg_print(NAND_DBG_WARN
,
4247 "NAND Program fail in "
4248 "%s, Line %d, Function: %s, "
4251 __FILE__
, __LINE__
, __func__
,
4252 (int)pbt
[wBlockNum
]);
4254 MARK_BLOCK_AS_BAD(pbt
[wBlockNum
]);
4259 wResult
= GLOB_LLD_Read_Page_Main(g_pTempBuf
,
4260 dwOldBlockAddr
, 0, DeviceInfo
.wPagesPerBlock
);
4261 if (wResult
== FAIL
)
4264 if (wResult
!= FAIL
) {
4265 /* This is a bug. At this time, pbt[wBlockNum]
4266 is still the physical address of
4267 discard block, and should not be write.
4268 Have fixed it as below.
4269 -- Yunpeng 2008.12.19
4271 wResult
= GLOB_LLD_Write_Page_Main(g_pTempBuf
,
4273 DeviceInfo
.wPagesPerBlock
);
4274 if (wResult
== FAIL
) {
4275 nand_dbg_print(NAND_DBG_WARN
,
4276 "NAND Program fail in "
4277 "%s, Line %d, Function: %s, "
4280 __FILE__
, __LINE__
, __func__
,
4281 (int)wReplacedNode
);
4282 MARK_BLOCK_AS_BAD(wReplacedNode
);
4284 pbt
[wBlockNum
] = wReplacedNode
;
4285 pbt
[wLeastReadIndex
] = wTempNode
;
4289 if ((wResult
== PASS
) && (g_cBlockTableStatus
!=
4290 IN_PROGRESS_BLOCK_TABLE
)) {
4291 g_cBlockTableStatus
= IN_PROGRESS_BLOCK_TABLE
;
4292 FTL_Write_IN_Progress_Block_Table_Page();
4296 } while (wResult
!= PASS
)