| Line 217... |
Line 217... |
| 217 |
__asm__ volatile(" stmfd sp!, {r1-r11,lr} \n");
|
217 |
__asm__ volatile(" stmfd sp!, {r1-r11,lr} \n");
|
| 218 |
__asm__ volatile(" mov r12, #0 \n");
|
218 |
__asm__ volatile(" mov r12, #0 \n");
|
| 219 |
__asm__ volatile(" str r12, [sp] \n");
|
219 |
__asm__ volatile(" str r12, [sp] \n");
|
| 220 |
__asm__ volatile(" mov r12, r2 \n");
|
220 |
__asm__ volatile(" mov r12, r2 \n");
|
| 221 |
__asm__ volatile(" add r8, r2, r2,lsl#1 \n");
|
221 |
__asm__ volatile(" add r8, r2, r2,lsl#1 \n");
|
| - |
|
222 |
__asm__ volatile(" add r8, r8, #3 \n");
|
| 222 |
__asm__ volatile(" add r3, r1, r3 \n");
|
223 |
__asm__ volatile(" add r3, r1, r3 \n");
|
| 223 |
__asm__ volatile(" sub r3, r3, #1 \n");
|
224 |
__asm__ volatile(" sub r3, r3, #1 \n");
|
| 224 |
__asm__ volatile(" mov r2, r1 \n");
|
225 |
__asm__ volatile(" mov r2, r1 \n");
|
| 225 |
__asm__ volatile(" add r1, r0, r12 \n");
|
226 |
__asm__ volatile(" add r1, r0, r12 \n");
|
| 226 |
__asm__ volatile(" sub r1, r1, #1 \n");
|
227 |
__asm__ volatile(" sub r1, r1, #1 \n");
|
| Line 256... |
Line 257... |
| 256 |
__asm__ volatile(" mov r5, #0 \n");
|
257 |
__asm__ volatile(" mov r5, #0 \n");
|
| 257 |
__asm__ volatile(" mov r6, #0 \n");
|
258 |
__asm__ volatile(" mov r6, #0 \n");
|
| 258 |
__asm__ volatile(" mov r7, r8 \n");
|
259 |
__asm__ volatile(" mov r7, r8 \n");
|
| 259 |
__asm__ volatile("displaylcd_dither_x: \n");
|
260 |
__asm__ volatile("displaylcd_dither_x: \n");
|
| 260 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
261 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 261 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
262 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 262 |
__asm__ volatile(" add r1, r1, r4 \n");
|
263 |
__asm__ volatile(" add r1, r1, r4 \n");
|
| 263 |
__asm__ volatile(" add r1, r1, r0 \n");
|
264 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 264 |
__asm__ volatile(" cmp r1, #0xff \n");
|
265 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 265 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
266 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 266 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
267 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
| Line 271... |
Line 272... |
| 271 |
__asm__ volatile(" mov r4, r4,lsr#1 \n");
|
272 |
__asm__ volatile(" mov r4, r4,lsr#1 \n");
|
| 272 |
__asm__ volatile(" add r4, r4, r1,lsr#2 \n");
|
273 |
__asm__ volatile(" add r4, r4, r1,lsr#2 \n");
|
| 273 |
__asm__ volatile(" strb r4, [r7], #1 \n");
|
274 |
__asm__ volatile(" strb r4, [r7], #1 \n");
|
| 274 |
__asm__ volatile(" mov r4, r1,asr#1 \n");
|
275 |
__asm__ volatile(" mov r4, r1,asr#1 \n");
|
| 275 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
276 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 276 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
277 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 277 |
__asm__ volatile(" add r1, r1, r5 \n");
|
278 |
__asm__ volatile(" add r1, r1, r5 \n");
|
| 278 |
__asm__ volatile(" add r1, r1, r0 \n");
|
279 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 279 |
__asm__ volatile(" cmp r1, #0xff \n");
|
280 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 280 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
281 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 281 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
282 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
| 282 |
__asm__ volatile(" mov r0, r1,lsr#2 \n");
|
283 |
__asm__ volatile(" mov r0, r1,lsr#2 \n");
|
| 283 |
__asm__ volatile(" orr r2, r2, r0,lsl#5 \n");
|
284 |
__asm__ volatile(" orr r2, r2, r0,lsl#5 \n");
|
| 284 |
__asm__ volatile(" sub r1, r1, r0,lsl#2 \n");
|
285 |
__asm__ volatile(" sub r1, r1, r0,lsl#2 \n");
|
| 285 |
__asm__ volatile(" sub r1, r1, r0,lsr#4 \n");
|
286 |
__asm__ volatile(" sub r1, r1, r0,lsr#4 \n");
|
| 286 |
__asm__ volatile(" mov r5, r5,lsr#1 \n");
|
287 |
__asm__ volatile(" mov r5, r5,lsr#1 \n");
|
| 287 |
__asm__ volatile(" add r5, r5, r1,lsr#2 \n");
|
288 |
__asm__ volatile(" add r5, r5, r1,lsr#2 \n");
|
| 288 |
__asm__ volatile(" strb r5, [r7], #1 \n");
|
289 |
__asm__ volatile(" strb r5, [r7], #1 \n");
|
| 289 |
__asm__ volatile(" mov r5, r1,asr#1 \n");
|
290 |
__asm__ volatile(" mov r5, r1,asr#1 \n");
|
| 290 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
291 |
__asm__ volatile(" ldrb r1, [r3], #1 \n");
|
| 291 |
__asm__ volatile(" ldrsb r0, [r7] \n");
|
292 |
__asm__ volatile(" ldrsb r0, [r7,#3] \n");
|
| 292 |
__asm__ volatile(" add r1, r1, r6 \n");
|
293 |
__asm__ volatile(" add r1, r1, r6 \n");
|
| 293 |
__asm__ volatile(" add r1, r1, r0 \n");
|
294 |
__asm__ volatile(" add r1, r1, r0 \n");
|
| 294 |
__asm__ volatile(" cmp r1, #0xff \n");
|
295 |
__asm__ volatile(" cmp r1, #0xff \n");
|
| 295 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
296 |
__asm__ volatile(" mvnhi r1, r1,asr#31 \n");
|
| 296 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|
297 |
__asm__ volatile(" andhi r1, r1, #0xff \n");
|