| Line 234... |
Line 234... |
| 234 |
__asm__ volatile(" stmfd sp!, {r1-r11,lr} \n");
|
234 |
__asm__ volatile(" stmfd sp!, {r1-r11,lr} \n");
|
| 235 |
__asm__ volatile(" mov r12, #0 \n");
|
235 |
__asm__ volatile(" mov r12, #0 \n");
|
| 236 |
__asm__ volatile(" str r12, [sp] \n");
|
236 |
__asm__ volatile(" str r12, [sp] \n");
|
| 237 |
__asm__ volatile(" mov r12, r2 \n");
|
237 |
__asm__ volatile(" mov r12, r2 \n");
|
| 238 |
__asm__ volatile(" add r8, r2, r2,lsl#1 \n");
|
238 |
__asm__ volatile(" add r8, r2, r2,lsl#1 \n");
|
| - |
|
239 |
__asm__ volatile(" add r8, r8, #3 \n");
|
| 239 |
__asm__ volatile(" add r3, r1, r3 \n");
|
240 |
__asm__ volatile(" add r3, r1, r3 \n");
|
| 240 |
__asm__ volatile(" sub r3, r3, #1 \n");
|
241 |
__asm__ volatile(" sub r3, r3, #1 \n");
|
| 241 |
__asm__ volatile(" mov r2, r1 \n");
|
242 |
__asm__ volatile(" mov r2, r1 \n");
|
| 242 |
__asm__ volatile(" add r1, r0, r12 \n");
|
243 |
__asm__ volatile(" add r1, r0, r12 \n");
|
| 243 |
__asm__ volatile(" sub r1, r1, #1 \n");
|
244 |
__asm__ volatile(" sub r1, r1, #1 \n");
|
| Line 279... |
Line 280... |
| 279 |
__asm__ volatile(" mov r5, #0 \n");
|
280 |
__asm__ volatile(" mov r5, #0 \n");
|
| 280 |
__asm__ volatile(" mov r6, #0 \n");
|
281 |
__asm__ volatile(" mov r6, #0 \n");
|
| 281 |
__asm__ volatile(" mov r7, r8 \n");
|
282 |
__asm__ volatile(" mov r7, r8 \n");
|
| 282 |
__asm__ volatile("displaylcd_dither_x: \n");
|
283 |
__asm__ volatile("displaylcd_dither_x: \n");
|
| 283 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
284 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 284 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
285 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 285 |
__asm__ volatile(" add r1, r1, r4 \n");
|
286 |
__asm__ volatile(" add r1, r1, r4 \n");
|
| 286 |
__asm__ volatile(" add r1, r1, r0 \n");
|
287 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 287 |
__asm__ volatile(" cmp r1, #0xff \n");
|
288 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 288 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
289 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 289 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
290 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
| Line 294... |
Line 295... |
| 294 |
__asm__ volatile(" mov r4, r4,lsr#1 \n");
|
295 |
__asm__ volatile(" mov r4, r4,lsr#1 \n");
|
| 295 |
__asm__ volatile(" add r4, r4, r1,lsr#2 \n");
|
296 |
__asm__ volatile(" add r4, r4, r1,lsr#2 \n");
|
| 296 |
__asm__ volatile(" strb r4, [r7], #1 \n");
|
297 |
__asm__ volatile(" strb r4, [r7], #1 \n");
|
| 297 |
__asm__ volatile(" mov r4, r1,asr#1 \n");
|
298 |
__asm__ volatile(" mov r4, r1,asr#1 \n");
|
| 298 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
299 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 299 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
300 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 300 |
__asm__ volatile(" add r1, r1, r5 \n");
|
301 |
__asm__ volatile(" add r1, r1, r5 \n");
|
| 301 |
__asm__ volatile(" add r1, r1, r0 \n");
|
302 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 302 |
__asm__ volatile(" cmp r1, #0xff \n");
|
303 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 303 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
304 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 304 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
305 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
| Line 309... |
Line 310... |
| 309 |
__asm__ volatile(" mov r5, r5,lsr#1 \n");
|
310 |
__asm__ volatile(" mov r5, r5,lsr#1 \n");
|
| 310 |
__asm__ volatile(" add r5, r5, r1,lsr#2 \n");
|
311 |
__asm__ volatile(" add r5, r5, r1,lsr#2 \n");
|
| 311 |
__asm__ volatile(" strb r5, [r7], #1 \n");
|
312 |
__asm__ volatile(" strb r5, [r7], #1 \n");
|
| 312 |
__asm__ volatile(" mov r5, r1,asr#1 \n");
|
313 |
__asm__ volatile(" mov r5, r1,asr#1 \n");
|
| 313 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
314 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 314 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
315 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 315 |
__asm__ volatile(" add r1, r1, r6 \n");
|
316 |
__asm__ volatile(" add r1, r1, r6 \n");
|
| 316 |
__asm__ volatile(" add r1, r1, r0 \n");
|
317 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 317 |
__asm__ volatile(" cmp r1, #0xff \n");
|
318 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 318 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
319 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 319 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
320 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|