diff options
Diffstat (limited to 'firmware/asm/arm')
-rw-r--r-- | firmware/asm/arm/lcd-as-memframe.S | 591 |
1 files changed, 0 insertions, 591 deletions
diff --git a/firmware/asm/arm/lcd-as-memframe.S b/firmware/asm/arm/lcd-as-memframe.S index 52ab0447c2..4bbae6fc0a 100644 --- a/firmware/asm/arm/lcd-as-memframe.S +++ b/firmware/asm/arm/lcd-as-memframe.S | |||
@@ -99,594 +99,3 @@ lcd_copy_buffer_rect: @ | |||
99 | bgt 10b @ copy line @ | 99 | bgt 10b @ copy line @ |
100 | ldmpc regs=r4-r11 @ restore regs and return | 100 | ldmpc regs=r4-r11 @ restore regs and return |
101 | .size lcd_copy_buffer_rect, .-lcd_copy_buffer_rect | 101 | .size lcd_copy_buffer_rect, .-lcd_copy_buffer_rect |
102 | |||
103 | |||
104 | /**************************************************************************** | ||
105 | * void lcd_write_yuv420_lines(fb_data *dst, | ||
106 | * unsigned char const * const src[3], | ||
107 | * int width, | ||
108 | * int stride); | ||
109 | * | ||
110 | * |R| |1.000000 -0.000001 1.402000| |Y'| | ||
111 | * |G| = |1.000000 -0.334136 -0.714136| |Pb| | ||
112 | * |B| |1.000000 1.772000 0.000000| |Pr| | ||
113 | * Scaled, normalized, rounded and tweaked to yield RGB 565: | ||
114 | * |R| |74 0 101| |Y' - 16| >> 9 | ||
115 | * |G| = |74 -24 -51| |Cb - 128| >> 8 | ||
116 | * |B| |74 128 0| |Cr - 128| >> 9 | ||
117 | * | ||
118 | * Write four RGB565 pixels in the following order on each loop: | ||
119 | * 1 3 + > down | ||
120 | * 2 4 \/ left | ||
121 | */ | ||
122 | .section .icode.lcd_write_yuv420_lines, "ax", %progbits | ||
123 | .align 2 | ||
124 | .global lcd_write_yuv420_lines | ||
125 | .type lcd_write_yuv420_lines, %function | ||
126 | lcd_write_yuv420_lines: | ||
127 | @ r0 = dst | ||
128 | @ r1 = yuv_src | ||
129 | @ r2 = width | ||
130 | @ r3 = stride | ||
131 | stmfd sp!, { r4-r10, lr } @ save non-scratch | ||
132 | ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p | ||
133 | @ r5 = yuv_src[1] = Cb_p | ||
134 | @ r6 = yuv_src[2] = Cr_p | ||
135 | @ r1 = scratch | ||
136 | sub r3, r3, #1 @ | ||
137 | 10: @ loop line @ | ||
138 | ldrb r7, [r4], #1 @ r7 = *Y'_p++; | ||
139 | ldrb r8, [r5], #1 @ r8 = *Cb_p++; | ||
140 | ldrb r9, [r6], #1 @ r9 = *Cr_p++; | ||
141 | @ | ||
142 | sub r7, r7, #16 @ r7 = Y = (Y' - 16)*74 | ||
143 | add r12, r7, r7, asl #2 @ actually (Y' - 16)*37 and shift right | ||
144 | add r7, r12, r7, asl #5 @ by one less when adding - same for all | ||
145 | @ | ||
146 | sub r8, r8, #128 @ Cb -= 128 | ||
147 | sub r9, r9, #128 @ Cr -= 128 | ||
148 | @ | ||
149 | add r10, r9, r9, asl #1 @ r10 = Cr*51 + Cb*24 | ||
150 | add r10, r10, r10, asl #4 @ | ||
151 | add r10, r10, r8, asl #3 @ | ||
152 | add r10, r10, r8, asl #4 @ | ||
153 | @ | ||
154 | add lr, r9, r9, asl #2 @ r9 = Cr*101 | ||
155 | add lr, lr, r9, asl #5 @ | ||
156 | add r9, lr, r9, asl #6 @ | ||
157 | @ | ||
158 | add r8, r8, #2 @ r8 = bu = (Cb*128 + 128) >> 8 | ||
159 | mov r8, r8, asr #2 @ | ||
160 | add r9, r9, #256 @ r9 = rv = (r9 + 256) >> 9 | ||
161 | mov r9, r9, asr #9 @ | ||
162 | rsb r10, r10, #128 @ r10 = guv = (-r10 + 128) >> 8 | ||
163 | mov r10, r10, asr #8 @ | ||
164 | @ compute R, G, and B | ||
165 | add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu | ||
166 | add lr, r9, r7, asr #8 @ lr = r = (Y >> 9) + rv | ||
167 | add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv | ||
168 | @ | ||
169 | #if ARM_ARCH >= 6 | ||
170 | usat r1, #5, r1 @ clamp b | ||
171 | usat lr, #5, lr @ clamp r | ||
172 | usat r7, #6, r7 @ clamp g | ||
173 | #else | ||
174 | orr r12, r1, lr @ check if clamping is needed... | ||
175 | orr r12, r12, r7, asr #1 @ ...at all | ||
176 | cmp r12, #31 @ | ||
177 | bls 15f @ no clamp @ | ||
178 | cmp r1, #31 @ clamp b | ||
179 | mvnhi r1, r1, asr #31 @ | ||
180 | andhi r1, r1, #31 @ | ||
181 | cmp lr, #31 @ clamp r | ||
182 | mvnhi lr, lr, asr #31 @ | ||
183 | andhi lr, lr, #31 @ | ||
184 | cmp r7, #63 @ clamp g | ||
185 | mvnhi r7, r7, asr #31 @ | ||
186 | andhi r7, r7, #63 @ | ||
187 | 15: @ no clamp @ | ||
188 | #endif | ||
189 | @ | ||
190 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
191 | @ | ||
192 | orr r1, r1, r7, lsl #5 @ r4 |= (g << 5) | ||
193 | orr r1, r1, lr, lsl #11 @ r4 = b | (r << 11) | ||
194 | |||
195 | #if LCD_WIDTH >= LCD_HEIGHT | ||
196 | strh r1, [r0] @ | ||
197 | #elif LCD_WIDTH < 256 | ||
198 | strh r1, [r0], #LCD_WIDTH @ store pixel | ||
199 | #else | ||
200 | strh r1, [r0] @ | ||
201 | #endif | ||
202 | @ | ||
203 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74 | ||
204 | add r12, r7, r7, asl #2 @ | ||
205 | add r7, r12, r7, asl #5 @ | ||
206 | @ compute R, G, and B | ||
207 | add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu | ||
208 | add lr, r9, r7, asr #8 @ lr = r = (Y >> 9) + rv | ||
209 | add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv | ||
210 | @ | ||
211 | #if ARM_ARCH >= 6 | ||
212 | usat r1, #5, r1 @ clamp b | ||
213 | usat lr, #5, lr @ clamp r | ||
214 | usat r7, #6, r7 @ clamp g | ||
215 | #else | ||
216 | orr r12, r1, lr @ check if clamping is needed... | ||
217 | orr r12, r12, r7, asr #1 @ ...at all | ||
218 | cmp r12, #31 @ | ||
219 | bls 15f @ no clamp @ | ||
220 | cmp r1, #31 @ clamp b | ||
221 | mvnhi r1, r1, asr #31 @ | ||
222 | andhi r1, r1, #31 @ | ||
223 | cmp lr, #31 @ clamp r | ||
224 | mvnhi lr, lr, asr #31 @ | ||
225 | andhi lr, lr, #31 @ | ||
226 | cmp r7, #63 @ clamp g | ||
227 | mvnhi r7, r7, asr #31 @ | ||
228 | andhi r7, r7, #63 @ | ||
229 | 15: @ no clamp @ | ||
230 | #endif | ||
231 | @ | ||
232 | ldrb r12, [r4], #1 @ r12 = Y' = *(Y'_p++) | ||
233 | @ | ||
234 | orr r1, r1, lr, lsl #11 @ r1 = b | (r << 11) | ||
235 | orr r1, r1, r7, lsl #5 @ r1 |= (g << 5) | ||
236 | |||
237 | #if LCD_WIDTH >= LCD_HEIGHT | ||
238 | add r0, r0, #2*LCD_WIDTH @ | ||
239 | strh r1, [r0] @ store pixel | ||
240 | sub r0, r0, #2*LCD_WIDTH @ | ||
241 | #elif LCD_WIDTH < 256 | ||
242 | strh r1, [r0, #-LCD_WIDTH-2] @ store pixel | ||
243 | #else | ||
244 | strh r1, [r0, #-2] @ | ||
245 | add r0, r0, #LCD_WIDTH @ | ||
246 | #endif | ||
247 | @ | ||
248 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74 | ||
249 | add r12, r7, r7, asl #2 @ | ||
250 | add r7, r12, r7, asl #5 @ | ||
251 | @ compute R, G, and B | ||
252 | add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu | ||
253 | add lr, r9, r7, asr #8 @ lr = r = (Y >> 9) + rv | ||
254 | add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv | ||
255 | @ | ||
256 | #if ARM_ARCH >= 6 | ||
257 | usat r1, #5, r1 @ clamp b | ||
258 | usat lr, #5, lr @ clamp r | ||
259 | usat r7, #6, r7 @ clamp g | ||
260 | #else | ||
261 | orr r12, r1, lr @ check if clamping is needed... | ||
262 | orr r12, r12, r7, asr #1 @ ...at all | ||
263 | cmp r12, #31 @ | ||
264 | bls 15f @ no clamp @ | ||
265 | cmp r1, #31 @ clamp b | ||
266 | mvnhi r1, r1, asr #31 @ | ||
267 | andhi r1, r1, #31 @ | ||
268 | cmp lr, #31 @ clamp r | ||
269 | mvnhi lr, lr, asr #31 @ | ||
270 | andhi lr, lr, #31 @ | ||
271 | cmp r7, #63 @ clamp g | ||
272 | mvnhi r7, r7, asr #31 @ | ||
273 | andhi r7, r7, #63 @ | ||
274 | 15: @ no clamp @ | ||
275 | #endif | ||
276 | @ | ||
277 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
278 | @ | ||
279 | orr r1, r1, r7, lsl #5 @ r1 = b | (g << 5) | ||
280 | orr r1, r1, lr, lsl #11 @ r1 |= (r << 11) | ||
281 | |||
282 | #if LCD_WIDTH >= LCD_HEIGHT | ||
283 | strh r1, [r0, #2] | ||
284 | #elif LCD_WIDTH < 256 | ||
285 | strh r1, [r0, #LCD_WIDTH]! @ store pixel | ||
286 | #else | ||
287 | strh r1, [r0] @ | ||
288 | #endif | ||
289 | @ | ||
290 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*74 | ||
291 | add r12, r7, r7, asl #2 @ | ||
292 | add r7, r12, r7, asl #5 @ | ||
293 | @ compute R, G, and B | ||
294 | add r1, r8, r7, asr #8 @ r1 = b = (Y >> 9) + bu | ||
295 | add lr, r9, r7, asr #8 @ lr = r = (Y >> 9) + rv | ||
296 | add r7, r10, r7, asr #7 @ r7 = g = (Y >> 8) + guv | ||
297 | @ | ||
298 | #if ARM_ARCH >= 6 | ||
299 | usat r1, #5, r1 @ clamp b | ||
300 | usat lr, #5, lr @ clamp r | ||
301 | usat r7, #6, r7 @ clamp g | ||
302 | #else | ||
303 | orr r12, r1, lr @ check if clamping is needed... | ||
304 | orr r12, r12, r7, asr #1 @ ...at all | ||
305 | cmp r12, #31 @ | ||
306 | bls 15f @ no clamp @ | ||
307 | cmp r1, #31 @ clamp b | ||
308 | mvnhi r1, r1, asr #31 @ | ||
309 | andhi r1, r1, #31 @ | ||
310 | cmp lr, #31 @ clamp r | ||
311 | mvnhi lr, lr, asr #31 @ | ||
312 | andhi lr, lr, #31 @ | ||
313 | cmp r7, #63 @ clamp g | ||
314 | mvnhi r7, r7, asr #31 @ | ||
315 | andhi r7, r7, #63 @ | ||
316 | 15: @ no clamp @ | ||
317 | #endif | ||
318 | @ | ||
319 | orr r12, r1, lr, lsl #11 @ r12 = b | (r << 11) | ||
320 | orr r12, r12, r7, lsl #5 @ r12 |= (g << 5) | ||
321 | |||
322 | #if LCD_WIDTH >= LCD_HEIGHT | ||
323 | add r0, r0, #2*LCD_WIDTH | ||
324 | strh r12, [r0, #2] | ||
325 | #if LCD_WIDTH <= 512 | ||
326 | sub r0, r0, #(2*LCD_WIDTH)-4 | ||
327 | #else | ||
328 | sub r0, r0, #(2*LCD_WIDTH) | ||
329 | add r0, r0, #4 | ||
330 | #endif | ||
331 | #else | ||
332 | strh r12, [r0, #-2] @ store pixel | ||
333 | #if LCD_WIDTH < 256 | ||
334 | add r0, r0, #2*LCD_WIDTH @ | ||
335 | #else | ||
336 | add r0, r0, #LCD_WIDTH @ | ||
337 | #endif | ||
338 | #endif | ||
339 | @ | ||
340 | subs r2, r2, #2 @ subtract block from width | ||
341 | bgt 10b @ loop line @ | ||
342 | @ | ||
343 | ldmpc regs=r4-r10 @ restore registers and return | ||
344 | .ltorg @ dump constant pool | ||
345 | .size lcd_write_yuv420_lines, .-lcd_write_yuv420_lines | ||
346 | |||
347 | |||
348 | /**************************************************************************** | ||
349 | * void lcd_write_yuv420_lines_odither(fb_data *dst, | ||
350 | * unsigned char const * const src[3], | ||
351 | * int width, | ||
352 | * int stride, | ||
353 | * int x_screen, | ||
354 | * int y_screen); | ||
355 | * | ||
356 | * |R| |1.000000 -0.000001 1.402000| |Y'| | ||
357 | * |G| = |1.000000 -0.334136 -0.714136| |Pb| | ||
358 | * |B| |1.000000 1.772000 0.000000| |Pr| | ||
359 | * Red scaled at twice g & b but at same precision to place it in correct | ||
360 | * bit position after multiply and leave instruction count lower. | ||
361 | * |R| |258 0 408| |Y' - 16| | ||
362 | * |G| = |149 -49 -104| |Cb - 128| | ||
363 | * |B| |149 258 0| |Cr - 128| | ||
364 | * | ||
365 | * Write four RGB565 pixels in the following order on each loop: | ||
366 | * 1 3 + > down | ||
367 | * 2 4 \/ left | ||
368 | * | ||
369 | * Kernel pattern (raw|rotated|use order): | ||
370 | * 5 3 4 2 2 6 3 7 row0 row2 > down | ||
371 | * 1 7 0 6 | 4 0 5 1 | 2 4 6 0 3 5 7 1 col0 left | ||
372 | * 4 2 5 3 | 3 7 2 6 | 3 5 7 1 2 4 6 0 col2 \/ | ||
373 | * 0 6 1 7 5 1 4 0 | ||
374 | */ | ||
375 | .section .icode.lcd_write_yuv420_lines_odither, "ax", %progbits | ||
376 | .align 2 | ||
377 | .global lcd_write_yuv420_lines_odither | ||
378 | .type lcd_write_yuv420_lines_odither, %function | ||
379 | lcd_write_yuv420_lines_odither: | ||
380 | @ r0 = dst | ||
381 | @ r1 = yuv_src | ||
382 | @ r2 = width | ||
383 | @ r3 = stride | ||
384 | @ [sp] = x_screen | ||
385 | @ [sp+4] = y_screen | ||
386 | stmfd sp!, { r4-r11, lr } @ save non-scratch | ||
387 | ldmia r1, { r4, r5, r6 } @ r4 = yuv_src[0] = Y'_p | ||
388 | @ r5 = yuv_src[1] = Cb_p | ||
389 | @ r6 = yuv_src[2] = Cr_p | ||
390 | @ | ||
391 | sub r3, r3, #1 @ | ||
392 | add r1, sp, #36 @ Line up pattern and kernel quadrant | ||
393 | ldmia r1, { r12, r14 } @ | ||
394 | eor r14, r14, r12 @ | ||
395 | and r14, r14, #0x2 @ | ||
396 | mov r14, r14, lsl #6 @ 0x00 or 0x80 | ||
397 | 10: @ loop line @ | ||
398 | @ | ||
399 | ldrb r7, [r4], #1 @ r7 = *Y'_p++; | ||
400 | ldrb r8, [r5], #1 @ r8 = *Cb_p++; | ||
401 | ldrb r9, [r6], #1 @ r9 = *Cr_p++; | ||
402 | @ | ||
403 | eor r14, r14, #0x80 @ flip pattern quadrant | ||
404 | @ | ||
405 | sub r7, r7, #16 @ r7 = Y = (Y' - 16)*149 | ||
406 | add r12, r7, r7, asl #2 @ | ||
407 | add r12, r12, r12, asl #4 @ | ||
408 | add r7, r12, r7, asl #6 @ | ||
409 | @ | ||
410 | sub r8, r8, #128 @ Cb -= 128 | ||
411 | sub r9, r9, #128 @ Cr -= 128 | ||
412 | @ | ||
413 | add r10, r8, r8, asl #4 @ r10 = guv = Cr*104 + Cb*49 | ||
414 | add r10, r10, r8, asl #5 @ | ||
415 | add r10, r10, r9, asl #3 @ | ||
416 | add r10, r10, r9, asl #5 @ | ||
417 | add r10, r10, r9, asl #6 @ | ||
418 | @ | ||
419 | mov r8, r8, asl #1 @ r8 = bu = Cb*258 | ||
420 | add r8, r8, r8, asl #7 @ | ||
421 | @ | ||
422 | add r9, r9, r9, asl #1 @ r9 = rv = Cr*408 | ||
423 | add r9, r9, r9, asl #4 @ | ||
424 | mov r9, r9, asl #3 @ | ||
425 | @ | ||
426 | @ compute R, G, and B | ||
427 | add r1, r8, r7 @ r1 = b' = Y + bu | ||
428 | add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv | ||
429 | rsb r7, r10, r7 @ r7 = g' = Y + guv | ||
430 | @ | ||
431 | @ r8 = bu, r9 = rv, r10 = guv | ||
432 | @ | ||
433 | sub r12, r1, r1, lsr #5 @ r1 = 31/32*b + b/256 | ||
434 | add r1, r12, r1, lsr #8 @ | ||
435 | @ | ||
436 | sub r12, r11, r11, lsr #5 @ r11 = 31/32*r + r/256 | ||
437 | add r11, r12, r11, lsr #8 @ | ||
438 | @ | ||
439 | sub r12, r7, r7, lsr #6 @ r7 = 63/64*g + g/256 | ||
440 | add r7, r12, r7, lsr #8 @ | ||
441 | @ | ||
442 | add r12, r14, #0x100 @ | ||
443 | @ | ||
444 | add r1, r1, r12 @ b = r1 + delta | ||
445 | add r11, r11, r12, lsl #1 @ r = r11 + delta*2 | ||
446 | add r7, r7, r12, lsr #1 @ g = r7 + delta/2 | ||
447 | @ | ||
448 | #if ARM_ARCH >= 6 | ||
449 | usat r11, #5, r11, asr #11 @ clamp r | ||
450 | usat r7, #6, r7, asr #9 @ clamp g | ||
451 | usat r1, #5, r1, asr #10 @ clamp b | ||
452 | @ | ||
453 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
454 | @ | ||
455 | orr r1, r1, r11, lsl #11 @ r1 = b | (r << 11) | ||
456 | orr r1, r1, r7, lsl #5 @ r1 |= (g << 5) | ||
457 | #else | ||
458 | orr r12, r1, r11, asr #1 @ check if clamping is needed... | ||
459 | orr r12, r12, r7 @ ...at all | ||
460 | movs r12, r12, asr #15 @ | ||
461 | beq 15f @ no clamp @ | ||
462 | movs r12, r1, asr #15 @ clamp b | ||
463 | mvnne r1, r12, lsr #15 @ | ||
464 | andne r1, r1, #0x7c00 @ mask b only if clamped | ||
465 | movs r12, r11, asr #16 @ clamp r | ||
466 | mvnne r11, r12, lsr #16 @ | ||
467 | movs r12, r7, asr #15 @ clamp g | ||
468 | mvnne r7, r12, lsr #15 @ | ||
469 | 15: @ no clamp @ | ||
470 | @ | ||
471 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
472 | @ | ||
473 | and r11, r11, #0xf800 @ pack pixel | ||
474 | and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) | | ||
475 | orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) | | ||
476 | orr r1, r11, r1, lsr #10 @ (b >> 10) | ||
477 | #endif | ||
478 | @ | ||
479 | #if LCD_WIDTH >= LCD_HEIGHT | ||
480 | strh r1, [r0] @ | ||
481 | #elif LCD_WIDTH < 256 | ||
482 | strh r1, [r0], #LCD_WIDTH @ store pixel | ||
483 | #else | ||
484 | strh r1, [r0] @ | ||
485 | #endif | ||
486 | @ | ||
487 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149 | ||
488 | add r12, r7, r7, asl #2 @ | ||
489 | add r12, r12, r12, asl #4 @ | ||
490 | add r7, r12, r7, asl #6 @ | ||
491 | @ compute R, G, and B | ||
492 | add r1, r8, r7 @ r1 = b' = Y + bu | ||
493 | add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv | ||
494 | rsb r7, r10, r7 @ r7 = g' = Y + guv | ||
495 | @ | ||
496 | sub r12, r1, r1, lsr #5 @ r1 = 31/32*b' + b'/256 | ||
497 | add r1, r12, r1, lsr #8 @ | ||
498 | @ | ||
499 | sub r12, r11, r11, lsr #5 @ r11 = 31/32*r' + r'/256 | ||
500 | add r11, r12, r11, lsr #8 @ | ||
501 | @ | ||
502 | sub r12, r7, r7, lsr #6 @ r7 = 63/64*g' + g'/256 | ||
503 | add r7, r12, r7, lsr #8 @ | ||
504 | @ | ||
505 | add r12, r14, #0x200 @ | ||
506 | @ | ||
507 | add r1, r1, r12 @ b = r1 + delta | ||
508 | add r11, r11, r12, lsl #1 @ r = r11 + delta*2 | ||
509 | add r7, r7, r12, lsr #1 @ g = r7 + delta/2 | ||
510 | @ | ||
511 | #if ARM_ARCH >= 6 | ||
512 | usat r11, #5, r11, asr #11 @ clamp r | ||
513 | usat r7, #6, r7, asr #9 @ clamp g | ||
514 | usat r1, #5, r1, asr #10 @ clamp b | ||
515 | @ | ||
516 | ldrb r12, [r4], #1 @ r12 = Y' = *(Y'_p++) | ||
517 | @ | ||
518 | orr r1, r1, r11, lsl #11 @ r1 = b | (r << 11) | ||
519 | orr r1, r1, r7, lsl #5 @ r1 |= (g << 5) | ||
520 | #else | ||
521 | orr r12, r1, r11, asr #1 @ check if clamping is needed... | ||
522 | orr r12, r12, r7 @ ...at all | ||
523 | movs r12, r12, asr #15 @ | ||
524 | beq 15f @ no clamp @ | ||
525 | movs r12, r1, asr #15 @ clamp b | ||
526 | mvnne r1, r12, lsr #15 @ | ||
527 | andne r1, r1, #0x7c00 @ mask b only if clamped | ||
528 | movs r12, r11, asr #16 @ clamp r | ||
529 | mvnne r11, r12, lsr #16 @ | ||
530 | movs r12, r7, asr #15 @ clamp g | ||
531 | mvnne r7, r12, lsr #15 @ | ||
532 | 15: @ no clamp @ | ||
533 | @ | ||
534 | ldrb r12, [r4], #1 @ r12 = Y' = *(Y'_p++) | ||
535 | @ | ||
536 | and r11, r11, #0xf800 @ pack pixel | ||
537 | and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) | | ||
538 | orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) | | ||
539 | orr r1, r11, r1, lsr #10 @ (b >> 10) | ||
540 | #endif | ||
541 | @ | ||
542 | #if LCD_WIDTH >= LCD_HEIGHT | ||
543 | add r0, r0, #2*LCD_WIDTH @ | ||
544 | strh r1, [r0] @ store pixel | ||
545 | sub r0, r0, #2*LCD_WIDTH @ | ||
546 | #elif LCD_WIDTH < 256 | ||
547 | strh r1, [r0, #-LCD_WIDTH-2] @ store pixel | ||
548 | #else | ||
549 | strh r1, [r0, #-2] @ store pixel | ||
550 | add r0, r0, #LCD_WIDTH @ | ||
551 | #endif | ||
552 | @ | ||
553 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149 | ||
554 | add r12, r7, r7, asl #2 @ | ||
555 | add r12, r12, r12, asl #4 @ | ||
556 | add r7, r12, r7, asl #6 @ | ||
557 | @ compute R, G, and B | ||
558 | add r1, r8, r7 @ r1 = b' = Y + bu | ||
559 | add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv | ||
560 | rsb r7, r10, r7 @ r7 = g' = Y + guv | ||
561 | @ | ||
562 | @ r8 = bu, r9 = rv, r10 = guv | ||
563 | @ | ||
564 | sub r12, r1, r1, lsr #5 @ r1 = 31/32*b' + b'/256 | ||
565 | add r1, r12, r1, lsr #8 @ | ||
566 | @ | ||
567 | sub r12, r11, r11, lsr #5 @ r11 = 31/32*r' + r'/256 | ||
568 | add r11, r12, r11, lsr #8 @ | ||
569 | @ | ||
570 | sub r12, r7, r7, lsr #6 @ r7 = 63/64*g' + g'/256 | ||
571 | add r7, r12, r7, lsr #8 @ | ||
572 | @ | ||
573 | add r12, r14, #0x300 @ | ||
574 | @ | ||
575 | add r1, r1, r12 @ b = r1 + delta | ||
576 | add r11, r11, r12, lsl #1 @ r = r11 + delta*2 | ||
577 | add r7, r7, r12, lsr #1 @ g = r7 + delta/2 | ||
578 | @ | ||
579 | #if ARM_ARCH >= 6 | ||
580 | usat r11, #5, r11, asr #11 @ clamp r | ||
581 | usat r7, #6, r7, asr #9 @ clamp g | ||
582 | usat r1, #5, r1, asr #10 @ clamp b | ||
583 | @ | ||
584 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
585 | @ | ||
586 | orr r1, r1, r11, lsl #11 @ r1 = b | (r << 11) | ||
587 | orr r1, r1, r7, lsl #5 @ r1 |= (g << 5) | ||
588 | #else | ||
589 | orr r12, r1, r11, asr #1 @ check if clamping is needed... | ||
590 | orr r12, r12, r7 @ ...at all | ||
591 | movs r12, r12, asr #15 @ | ||
592 | beq 15f @ no clamp @ | ||
593 | movs r12, r1, asr #15 @ clamp b | ||
594 | mvnne r1, r12, lsr #15 @ | ||
595 | andne r1, r1, #0x7c00 @ mask b only if clamped | ||
596 | movs r12, r11, asr #16 @ clamp r | ||
597 | mvnne r11, r12, lsr #16 @ | ||
598 | movs r12, r7, asr #15 @ clamp g | ||
599 | mvnne r7, r12, lsr #15 @ | ||
600 | 15: @ no clamp @ | ||
601 | @ | ||
602 | ldrb r12, [r4, r3] @ r12 = Y' = *(Y'_p + stride) | ||
603 | @ | ||
604 | and r11, r11, #0xf800 @ pack pixel | ||
605 | and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) | | ||
606 | orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) | | ||
607 | orr r1, r11, r1, lsr #10 @ (b >> 10) | ||
608 | #endif | ||
609 | @ | ||
610 | #if LCD_WIDTH >= LCD_HEIGHT | ||
611 | strh r1, [r0, #2] | ||
612 | #elif LCD_WIDTH < 256 | ||
613 | strh r1, [r0, #LCD_WIDTH]! @ store pixel | ||
614 | #else | ||
615 | strh r1, [r0] @ | ||
616 | #endif | ||
617 | |||
618 | sub r7, r12, #16 @ r7 = Y = (Y' - 16)*149 | ||
619 | add r12, r7, r7, asl #2 @ | ||
620 | add r12, r12, r12, asl #4 @ | ||
621 | add r7, r12, r7, asl #6 @ | ||
622 | @ compute R, G, and B | ||
623 | add r1, r8, r7 @ r1 = b' = Y + bu | ||
624 | add r11, r9, r7, asl #1 @ r11 = r' = Y*2 + rv | ||
625 | rsb r7, r10, r7 @ r7 = g' = Y + guv | ||
626 | @ | ||
627 | sub r12, r1, r1, lsr #5 @ r1 = 31/32*b + b/256 | ||
628 | add r1, r12, r1, lsr #8 @ | ||
629 | @ | ||
630 | sub r12, r11, r11, lsr #5 @ r11 = 31/32*r + r/256 | ||
631 | add r11, r12, r11, lsr #8 @ | ||
632 | @ | ||
633 | sub r12, r7, r7, lsr #6 @ r7 = 63/64*g + g/256 | ||
634 | add r7, r12, r7, lsr #8 @ | ||
635 | @ | ||
636 | @ This element is zero - use r14 @ | ||
637 | @ | ||
638 | add r1, r1, r14 @ b = r1 + delta | ||
639 | add r11, r11, r14, lsl #1 @ r = r11 + delta*2 | ||
640 | add r7, r7, r14, lsr #1 @ g = r7 + delta/2 | ||
641 | @ | ||
642 | #if ARM_ARCH >= 6 | ||
643 | usat r11, #5, r11, asr #11 @ clamp r | ||
644 | usat r7, #6, r7, asr #9 @ clamp g | ||
645 | usat r1, #5, r1, asr #10 @ clamp b | ||
646 | @ | ||
647 | orr r1, r1, r11, lsl #11 @ r1 = b | (r << 11) | ||
648 | orr r1, r1, r7, lsl #5 @ r1 |= (g << 5) | ||
649 | #else | ||
650 | orr r12, r1, r11, asr #1 @ check if clamping is needed... | ||
651 | orr r12, r12, r7 @ ...at all | ||
652 | movs r12, r12, asr #15 @ | ||
653 | beq 15f @ no clamp @ | ||
654 | movs r12, r1, asr #15 @ clamp b | ||
655 | mvnne r1, r12, lsr #15 @ | ||
656 | andne r1, r1, #0x7c00 @ mask b only if clamped | ||
657 | movs r12, r11, asr #16 @ clamp r | ||
658 | mvnne r11, r12, lsr #16 @ | ||
659 | movs r12, r7, asr #15 @ clamp g | ||
660 | mvnne r7, r12, lsr #15 @ | ||
661 | 15: @ no clamp @ | ||
662 | @ | ||
663 | and r11, r11, #0xf800 @ pack pixel | ||
664 | and r7, r7, #0x7e00 @ r1 = pixel = (r & 0xf800) | | ||
665 | orr r11, r11, r7, lsr #4 @ ((g & 0x7e00) >> 4) | | ||
666 | orr r1, r11, r1, lsr #10 @ (b >> 10) | ||
667 | #endif | ||
668 | @ | ||
669 | #if LCD_WIDTH >= LCD_HEIGHT | ||
670 | add r0, r0, #2*LCD_WIDTH | ||
671 | strh r1, [r0, #2] @ store pixel | ||
672 | #if LCD_WIDTH <= 512 | ||
673 | sub r0, r0, #(2*LCD_WIDTH)-4 | ||
674 | #else | ||
675 | sub r0, r0, #(2*LCD_WIDTH) | ||
676 | add r0, r0, #4 | ||
677 | #endif | ||
678 | #else | ||
679 | strh r1, [r0, #-2] @ store pixel | ||
680 | #if LCD_WIDTH < 256 | ||
681 | add r0, r0, #2*LCD_WIDTH @ | ||
682 | #else | ||
683 | add r0, r0, #LCD_WIDTH @ | ||
684 | #endif | ||
685 | #endif | ||
686 | @ | ||
687 | subs r2, r2, #2 @ subtract block from width | ||
688 | bgt 10b @ loop line @ | ||
689 | @ | ||
690 | ldmpc regs=r4-r11 @ restore registers and return | ||
691 | .ltorg @ dump constant pool | ||
692 | .size lcd_write_yuv420_lines_odither, .-lcd_write_yuv420_lines_odither | ||