Blame view

drivers/dma/ipu/ipu_idmac.c 45.8 KB
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1
2
3
4
5
6
7
8
9
10
  /*
   * Copyright (C) 2008
   * Guennadi Liakhovetski, DENX Software Engineering, <lg@denx.de>
   *
   * Copyright (C) 2005-2007 Freescale Semiconductor, Inc. All Rights Reserved.
   *
   * This program is free software; you can redistribute it and/or modify
   * it under the terms of the GNU General Public License version 2 as
   * published by the Free Software Foundation.
   */
b7f080cfe   Alexey Dobriyan   net: remove mm.h ...
11
  #include <linux/dma-mapping.h>
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
  #include <linux/init.h>
  #include <linux/platform_device.h>
  #include <linux/err.h>
  #include <linux/spinlock.h>
  #include <linux/delay.h>
  #include <linux/list.h>
  #include <linux/clk.h>
  #include <linux/vmalloc.h>
  #include <linux/string.h>
  #include <linux/interrupt.h>
  #include <linux/io.h>
  
  #include <mach/ipu.h>
  
  #include "ipu_intern.h"
  
  #define FS_VF_IN_VALID	0x00000002
  #define FS_ENC_IN_VALID	0x00000001
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
30
31
  static int ipu_disable_channel(struct idmac *idmac, struct idmac_channel *ichan,
  			       bool wait_for_stop);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
  /*
   * There can be only one, we could allocate it dynamically, but then we'd have
   * to add an extra parameter to some functions, and use something as ugly as
   *	struct ipu *ipu = to_ipu(to_idmac(ichan->dma_chan.device));
   * in the ISR
   */
  static struct ipu ipu_data;
  
  #define to_ipu(id) container_of(id, struct ipu, idmac)
  
  static u32 __idmac_read_icreg(struct ipu *ipu, unsigned long reg)
  {
  	return __raw_readl(ipu->reg_ic + reg);
  }
  
  #define idmac_read_icreg(ipu, reg) __idmac_read_icreg(ipu, reg - IC_CONF)
  
  static void __idmac_write_icreg(struct ipu *ipu, u32 value, unsigned long reg)
  {
  	__raw_writel(value, ipu->reg_ic + reg);
  }
  
  #define idmac_write_icreg(ipu, v, reg) __idmac_write_icreg(ipu, v, reg - IC_CONF)
  
  static u32 idmac_read_ipureg(struct ipu *ipu, unsigned long reg)
  {
  	return __raw_readl(ipu->reg_ipu + reg);
  }
  
  static void idmac_write_ipureg(struct ipu *ipu, u32 value, unsigned long reg)
  {
  	__raw_writel(value, ipu->reg_ipu + reg);
  }
  
  /*****************************************************************************
   * IPU / IC common functions
   */
  static void dump_idmac_reg(struct ipu *ipu)
  {
  	dev_dbg(ipu->dev, "IDMAC_CONF 0x%x, IC_CONF 0x%x, IDMAC_CHA_EN 0x%x, "
  		"IDMAC_CHA_PRI 0x%x, IDMAC_CHA_BUSY 0x%x
  ",
  		idmac_read_icreg(ipu, IDMAC_CONF),
  		idmac_read_icreg(ipu, IC_CONF),
  		idmac_read_icreg(ipu, IDMAC_CHA_EN),
  		idmac_read_icreg(ipu, IDMAC_CHA_PRI),
  		idmac_read_icreg(ipu, IDMAC_CHA_BUSY));
  	dev_dbg(ipu->dev, "BUF0_RDY 0x%x, BUF1_RDY 0x%x, CUR_BUF 0x%x, "
  		"DB_MODE 0x%x, TASKS_STAT 0x%x
  ",
  		idmac_read_ipureg(ipu, IPU_CHA_BUF0_RDY),
  		idmac_read_ipureg(ipu, IPU_CHA_BUF1_RDY),
  		idmac_read_ipureg(ipu, IPU_CHA_CUR_BUF),
  		idmac_read_ipureg(ipu, IPU_CHA_DB_MODE_SEL),
  		idmac_read_ipureg(ipu, IPU_TASKS_STAT));
  }
  
  static uint32_t bytes_per_pixel(enum pixel_fmt fmt)
  {
  	switch (fmt) {
  	case IPU_PIX_FMT_GENERIC:	/* generic data */
  	case IPU_PIX_FMT_RGB332:
  	case IPU_PIX_FMT_YUV420P:
  	case IPU_PIX_FMT_YUV422P:
  	default:
  		return 1;
  	case IPU_PIX_FMT_RGB565:
  	case IPU_PIX_FMT_YUYV:
  	case IPU_PIX_FMT_UYVY:
  		return 2;
  	case IPU_PIX_FMT_BGR24:
  	case IPU_PIX_FMT_RGB24:
  		return 3;
  	case IPU_PIX_FMT_GENERIC_32:	/* generic data */
  	case IPU_PIX_FMT_BGR32:
  	case IPU_PIX_FMT_RGB32:
  	case IPU_PIX_FMT_ABGR32:
  		return 4;
  	}
  }
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
112
  /* Enable direct write to memory by the Camera Sensor Interface */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
  static void ipu_ic_enable_task(struct ipu *ipu, enum ipu_channel channel)
  {
  	uint32_t ic_conf, mask;
  
  	switch (channel) {
  	case IDMAC_IC_0:
  		mask = IC_CONF_PRPENC_EN;
  		break;
  	case IDMAC_IC_7:
  		mask = IC_CONF_RWS_EN | IC_CONF_PRPENC_EN;
  		break;
  	default:
  		return;
  	}
  	ic_conf = idmac_read_icreg(ipu, IC_CONF) | mask;
  	idmac_write_icreg(ipu, ic_conf, IC_CONF);
  }
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
130
  /* Called under spin_lock_irqsave(&ipu_data.lock) */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
  static void ipu_ic_disable_task(struct ipu *ipu, enum ipu_channel channel)
  {
  	uint32_t ic_conf, mask;
  
  	switch (channel) {
  	case IDMAC_IC_0:
  		mask = IC_CONF_PRPENC_EN;
  		break;
  	case IDMAC_IC_7:
  		mask = IC_CONF_RWS_EN | IC_CONF_PRPENC_EN;
  		break;
  	default:
  		return;
  	}
  	ic_conf = idmac_read_icreg(ipu, IC_CONF) & ~mask;
  	idmac_write_icreg(ipu, ic_conf, IC_CONF);
  }
  
  static uint32_t ipu_channel_status(struct ipu *ipu, enum ipu_channel channel)
  {
  	uint32_t stat = TASK_STAT_IDLE;
  	uint32_t task_stat_reg = idmac_read_ipureg(ipu, IPU_TASKS_STAT);
  
  	switch (channel) {
  	case IDMAC_IC_7:
  		stat = (task_stat_reg & TSTAT_CSI2MEM_MASK) >>
  			TSTAT_CSI2MEM_OFFSET;
  		break;
  	case IDMAC_IC_0:
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  	default:
  		break;
  	}
  	return stat;
  }
  
  struct chan_param_mem_planar {
  	/* Word 0 */
  	u32	xv:10;
  	u32	yv:10;
  	u32	xb:12;
  
  	u32	yb:12;
  	u32	res1:2;
  	u32	nsb:1;
  	u32	lnpb:6;
  	u32	ubo_l:11;
  
  	u32	ubo_h:15;
  	u32	vbo_l:17;
  
  	u32	vbo_h:9;
  	u32	res2:3;
  	u32	fw:12;
  	u32	fh_l:8;
  
  	u32	fh_h:4;
  	u32	res3:28;
  
  	/* Word 1 */
  	u32	eba0;
  
  	u32	eba1;
  
  	u32	bpp:3;
  	u32	sl:14;
  	u32	pfs:3;
  	u32	bam:3;
  	u32	res4:2;
  	u32	npb:6;
  	u32	res5:1;
  
  	u32	sat:2;
  	u32	res6:30;
  } __attribute__ ((packed));
  
  struct chan_param_mem_interleaved {
  	/* Word 0 */
  	u32	xv:10;
  	u32	yv:10;
  	u32	xb:12;
  
  	u32	yb:12;
  	u32	sce:1;
  	u32	res1:1;
  	u32	nsb:1;
  	u32	lnpb:6;
  	u32	sx:10;
  	u32	sy_l:1;
  
  	u32	sy_h:9;
  	u32	ns:10;
  	u32	sm:10;
  	u32	sdx_l:3;
  
  	u32	sdx_h:2;
  	u32	sdy:5;
  	u32	sdrx:1;
  	u32	sdry:1;
  	u32	sdr1:1;
  	u32	res2:2;
  	u32	fw:12;
  	u32	fh_l:8;
  
  	u32	fh_h:4;
  	u32	res3:28;
  
  	/* Word 1 */
  	u32	eba0;
  
  	u32	eba1;
  
  	u32	bpp:3;
  	u32	sl:14;
  	u32	pfs:3;
  	u32	bam:3;
  	u32	res4:2;
  	u32	npb:6;
  	u32	res5:1;
  
  	u32	sat:2;
  	u32	scc:1;
  	u32	ofs0:5;
  	u32	ofs1:5;
  	u32	ofs2:5;
  	u32	ofs3:5;
  	u32	wid0:3;
  	u32	wid1:3;
  	u32	wid2:3;
  
  	u32	wid3:3;
  	u32	dec_sel:1;
  	u32	res6:28;
  } __attribute__ ((packed));
  
  union chan_param_mem {
  	struct chan_param_mem_planar		pp;
  	struct chan_param_mem_interleaved	ip;
  };
  
  static void ipu_ch_param_set_plane_offset(union chan_param_mem *params,
  					  u32 u_offset, u32 v_offset)
  {
  	params->pp.ubo_l = u_offset & 0x7ff;
  	params->pp.ubo_h = u_offset >> 11;
  	params->pp.vbo_l = v_offset & 0x1ffff;
  	params->pp.vbo_h = v_offset >> 17;
  }
  
  static void ipu_ch_param_set_size(union chan_param_mem *params,
  				  uint32_t pixel_fmt, uint16_t width,
  				  uint16_t height, uint16_t stride)
  {
  	u32 u_offset;
  	u32 v_offset;
  
  	params->pp.fw		= width - 1;
  	params->pp.fh_l		= height - 1;
  	params->pp.fh_h		= (height - 1) >> 8;
  	params->pp.sl		= stride - 1;
  
  	switch (pixel_fmt) {
  	case IPU_PIX_FMT_GENERIC:
  		/*Represents 8-bit Generic data */
  		params->pp.bpp	= 3;
  		params->pp.pfs	= 7;
  		params->pp.npb	= 31;
  		params->pp.sat	= 2;		/* SAT = use 32-bit access */
  		break;
  	case IPU_PIX_FMT_GENERIC_32:
  		/*Represents 32-bit Generic data */
  		params->pp.bpp	= 0;
  		params->pp.pfs	= 7;
  		params->pp.npb	= 7;
  		params->pp.sat	= 2;		/* SAT = use 32-bit access */
  		break;
  	case IPU_PIX_FMT_RGB565:
  		params->ip.bpp	= 2;
  		params->ip.pfs	= 4;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		params->ip.ofs0	= 0;		/* Red bit offset */
  		params->ip.ofs1	= 5;		/* Green bit offset */
  		params->ip.ofs2	= 11;		/* Blue bit offset */
  		params->ip.ofs3	= 16;		/* Alpha bit offset */
  		params->ip.wid0	= 4;		/* Red bit width - 1 */
  		params->ip.wid1	= 5;		/* Green bit width - 1 */
  		params->ip.wid2	= 4;		/* Blue bit width - 1 */
  		break;
  	case IPU_PIX_FMT_BGR24:
  		params->ip.bpp	= 1;		/* 24 BPP & RGB PFS */
  		params->ip.pfs	= 4;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		params->ip.ofs0	= 0;		/* Red bit offset */
  		params->ip.ofs1	= 8;		/* Green bit offset */
  		params->ip.ofs2	= 16;		/* Blue bit offset */
  		params->ip.ofs3	= 24;		/* Alpha bit offset */
  		params->ip.wid0	= 7;		/* Red bit width - 1 */
  		params->ip.wid1	= 7;		/* Green bit width - 1 */
  		params->ip.wid2	= 7;		/* Blue bit width - 1 */
  		break;
  	case IPU_PIX_FMT_RGB24:
  		params->ip.bpp	= 1;		/* 24 BPP & RGB PFS */
  		params->ip.pfs	= 4;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		params->ip.ofs0	= 16;		/* Red bit offset */
  		params->ip.ofs1	= 8;		/* Green bit offset */
  		params->ip.ofs2	= 0;		/* Blue bit offset */
  		params->ip.ofs3	= 24;		/* Alpha bit offset */
  		params->ip.wid0	= 7;		/* Red bit width - 1 */
  		params->ip.wid1	= 7;		/* Green bit width - 1 */
  		params->ip.wid2	= 7;		/* Blue bit width - 1 */
  		break;
  	case IPU_PIX_FMT_BGRA32:
  	case IPU_PIX_FMT_BGR32:
9ad7bd294   Roel Kluin   dma: cases IPU_PI...
349
  	case IPU_PIX_FMT_ABGR32:
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
  		params->ip.bpp	= 0;
  		params->ip.pfs	= 4;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		params->ip.ofs0	= 8;		/* Red bit offset */
  		params->ip.ofs1	= 16;		/* Green bit offset */
  		params->ip.ofs2	= 24;		/* Blue bit offset */
  		params->ip.ofs3	= 0;		/* Alpha bit offset */
  		params->ip.wid0	= 7;		/* Red bit width - 1 */
  		params->ip.wid1	= 7;		/* Green bit width - 1 */
  		params->ip.wid2	= 7;		/* Blue bit width - 1 */
  		params->ip.wid3	= 7;		/* Alpha bit width - 1 */
  		break;
  	case IPU_PIX_FMT_RGBA32:
  	case IPU_PIX_FMT_RGB32:
  		params->ip.bpp	= 0;
  		params->ip.pfs	= 4;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		params->ip.ofs0	= 24;		/* Red bit offset */
  		params->ip.ofs1	= 16;		/* Green bit offset */
  		params->ip.ofs2	= 8;		/* Blue bit offset */
  		params->ip.ofs3	= 0;		/* Alpha bit offset */
  		params->ip.wid0	= 7;		/* Red bit width - 1 */
  		params->ip.wid1	= 7;		/* Green bit width - 1 */
  		params->ip.wid2	= 7;		/* Blue bit width - 1 */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
  		params->ip.wid3	= 7;		/* Alpha bit width - 1 */
  		break;
  	case IPU_PIX_FMT_UYVY:
  		params->ip.bpp	= 2;
  		params->ip.pfs	= 6;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		break;
  	case IPU_PIX_FMT_YUV420P2:
  	case IPU_PIX_FMT_YUV420P:
  		params->ip.bpp	= 3;
  		params->ip.pfs	= 3;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		u_offset = stride * height;
  		v_offset = u_offset + u_offset / 4;
  		ipu_ch_param_set_plane_offset(params, u_offset, v_offset);
  		break;
  	case IPU_PIX_FMT_YVU422P:
  		params->ip.bpp	= 3;
  		params->ip.pfs	= 2;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		v_offset = stride * height;
  		u_offset = v_offset + v_offset / 2;
  		ipu_ch_param_set_plane_offset(params, u_offset, v_offset);
  		break;
  	case IPU_PIX_FMT_YUV422P:
  		params->ip.bpp	= 3;
  		params->ip.pfs	= 2;
  		params->ip.npb	= 7;
  		params->ip.sat	= 2;		/* SAT = 32-bit access */
  		u_offset = stride * height;
  		v_offset = u_offset + u_offset / 2;
  		ipu_ch_param_set_plane_offset(params, u_offset, v_offset);
  		break;
  	default:
  		dev_err(ipu_data.dev,
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
414
415
  			"mx3 ipu: unimplemented pixel format %d
  ", pixel_fmt);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
416
417
418
419
420
421
422
423
424
425
  		break;
  	}
  
  	params->pp.nsb = 1;
  }
  
  static void ipu_ch_param_set_burst_size(union chan_param_mem *params,
  					uint16_t burst_pixels)
  {
  	params->pp.npb = burst_pixels - 1;
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
426
  }
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
427
428
429
430
431
432
  
  static void ipu_ch_param_set_buffer(union chan_param_mem *params,
  				    dma_addr_t buf0, dma_addr_t buf1)
  {
  	params->pp.eba0 = buf0;
  	params->pp.eba1 = buf1;
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
433
  }
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
434
435
436
437
438
  
  static void ipu_ch_param_set_rotation(union chan_param_mem *params,
  				      enum ipu_rotate_mode rotate)
  {
  	params->pp.bam = rotate;
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
439
  }
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
  
  static void ipu_write_param_mem(uint32_t addr, uint32_t *data,
  				uint32_t num_words)
  {
  	for (; num_words > 0; num_words--) {
  		dev_dbg(ipu_data.dev,
  			"write param mem - addr = 0x%08X, data = 0x%08X
  ",
  			addr, *data);
  		idmac_write_ipureg(&ipu_data, addr, IPU_IMA_ADDR);
  		idmac_write_ipureg(&ipu_data, *data++, IPU_IMA_DATA);
  		addr++;
  		if ((addr & 0x7) == 5) {
  			addr &= ~0x7;	/* set to word 0 */
  			addr += 8;	/* increment to next row */
  		}
  	}
  }
  
  static int calc_resize_coeffs(uint32_t in_size, uint32_t out_size,
  			      uint32_t *resize_coeff,
  			      uint32_t *downsize_coeff)
  {
  	uint32_t temp_size;
  	uint32_t temp_downsize;
  
  	*resize_coeff	= 1 << 13;
  	*downsize_coeff	= 1 << 13;
  
  	/* Cannot downsize more than 8:1 */
  	if (out_size << 3 < in_size)
  		return -EINVAL;
  
  	/* compute downsizing coefficient */
  	temp_downsize = 0;
  	temp_size = in_size;
  	while (temp_size >= out_size * 2 && temp_downsize < 2) {
  		temp_size >>= 1;
  		temp_downsize++;
  	}
  	*downsize_coeff = temp_downsize;
  
  	/*
  	 * compute resizing coefficient using the following formula:
  	 * resize_coeff = M*(SI -1)/(SO - 1)
  	 * where M = 2^13, SI - input size, SO - output size
  	 */
  	*resize_coeff = (8192L * (temp_size - 1)) / (out_size - 1);
  	if (*resize_coeff >= 16384L) {
  		dev_err(ipu_data.dev, "Warning! Overflow on resize coeff.
  ");
  		*resize_coeff = 0x3FFF;
  	}
  
  	dev_dbg(ipu_data.dev, "resizing from %u -> %u pixels, "
  		"downsize=%u, resize=%u.%lu (reg=%u)
  ", in_size, out_size,
  		*downsize_coeff, *resize_coeff >= 8192L ? 1 : 0,
  		((*resize_coeff & 0x1FFF) * 10000L) / 8192L, *resize_coeff);
  
  	return 0;
  }
  
  static enum ipu_color_space format_to_colorspace(enum pixel_fmt fmt)
  {
  	switch (fmt) {
  	case IPU_PIX_FMT_RGB565:
  	case IPU_PIX_FMT_BGR24:
  	case IPU_PIX_FMT_RGB24:
  	case IPU_PIX_FMT_BGR32:
  	case IPU_PIX_FMT_RGB32:
  		return IPU_COLORSPACE_RGB;
  	default:
  		return IPU_COLORSPACE_YCBCR;
  	}
  }
  
  static int ipu_ic_init_prpenc(struct ipu *ipu,
  			      union ipu_channel_param *params, bool src_is_csi)
  {
  	uint32_t reg, ic_conf;
  	uint32_t downsize_coeff, resize_coeff;
  	enum ipu_color_space in_fmt, out_fmt;
  
  	/* Setup vertical resizing */
  	calc_resize_coeffs(params->video.in_height,
  			    params->video.out_height,
  			    &resize_coeff, &downsize_coeff);
  	reg = (downsize_coeff << 30) | (resize_coeff << 16);
  
  	/* Setup horizontal resizing */
  	calc_resize_coeffs(params->video.in_width,
  			    params->video.out_width,
  			    &resize_coeff, &downsize_coeff);
  	reg |= (downsize_coeff << 14) | resize_coeff;
  
  	/* Setup color space conversion */
  	in_fmt = format_to_colorspace(params->video.in_pixel_fmt);
  	out_fmt = format_to_colorspace(params->video.out_pixel_fmt);
  
  	/*
  	 * Colourspace conversion unsupported yet - see _init_csc() in
  	 * Freescale sources
  	 */
  	if (in_fmt != out_fmt) {
  		dev_err(ipu->dev, "Colourspace conversion unsupported!
  ");
  		return -EOPNOTSUPP;
  	}
  
  	idmac_write_icreg(ipu, reg, IC_PRP_ENC_RSC);
  
  	ic_conf = idmac_read_icreg(ipu, IC_CONF);
  
  	if (src_is_csi)
  		ic_conf &= ~IC_CONF_RWS_EN;
  	else
  		ic_conf |= IC_CONF_RWS_EN;
  
  	idmac_write_icreg(ipu, ic_conf, IC_CONF);
  
  	return 0;
  }
  
  static uint32_t dma_param_addr(uint32_t dma_ch)
  {
  	/* Channel Parameter Memory */
  	return 0x10000 | (dma_ch << 4);
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
568
  }
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
  
  static void ipu_channel_set_priority(struct ipu *ipu, enum ipu_channel channel,
  				     bool prio)
  {
  	u32 reg = idmac_read_icreg(ipu, IDMAC_CHA_PRI);
  
  	if (prio)
  		reg |= 1UL << channel;
  	else
  		reg &= ~(1UL << channel);
  
  	idmac_write_icreg(ipu, reg, IDMAC_CHA_PRI);
  
  	dump_idmac_reg(ipu);
  }
  
  static uint32_t ipu_channel_conf_mask(enum ipu_channel channel)
  {
  	uint32_t mask;
  
  	switch (channel) {
  	case IDMAC_IC_0:
  	case IDMAC_IC_7:
  		mask = IPU_CONF_CSI_EN | IPU_CONF_IC_EN;
  		break;
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  		mask = IPU_CONF_SDC_EN | IPU_CONF_DI_EN;
  		break;
  	default:
  		mask = 0;
  		break;
  	}
  
  	return mask;
  }
  
  /**
   * ipu_enable_channel() - enable an IPU channel.
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
608
609
   * @idmac:	IPU DMAC context.
   * @ichan:	IDMAC channel.
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
   * @return:	0 on success or negative error code on failure.
   */
  static int ipu_enable_channel(struct idmac *idmac, struct idmac_channel *ichan)
  {
  	struct ipu *ipu = to_ipu(idmac);
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
  	uint32_t reg;
  	unsigned long flags;
  
  	spin_lock_irqsave(&ipu->lock, flags);
  
  	/* Reset to buffer 0 */
  	idmac_write_ipureg(ipu, 1UL << channel, IPU_CHA_CUR_BUF);
  	ichan->active_buffer = 0;
  	ichan->status = IPU_CHANNEL_ENABLED;
  
  	switch (channel) {
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  	case IDMAC_IC_7:
  		ipu_channel_set_priority(ipu, channel, true);
  	default:
  		break;
  	}
  
  	reg = idmac_read_icreg(ipu, IDMAC_CHA_EN);
  
  	idmac_write_icreg(ipu, reg | (1UL << channel), IDMAC_CHA_EN);
  
  	ipu_ic_enable_task(ipu, channel);
  
  	spin_unlock_irqrestore(&ipu->lock, flags);
  	return 0;
  }
  
  /**
   * ipu_init_channel_buffer() - initialize a buffer for logical IPU channel.
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
647
   * @ichan:	IDMAC channel.
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
   * @pixel_fmt:	pixel format of buffer. Pixel format is a FOURCC ASCII code.
   * @width:	width of buffer in pixels.
   * @height:	height of buffer in pixels.
   * @stride:	stride length of buffer in pixels.
   * @rot_mode:	rotation mode of buffer. A rotation setting other than
   *		IPU_ROTATE_VERT_FLIP should only be used for input buffers of
   *		rotation channels.
   * @phyaddr_0:	buffer 0 physical address.
   * @phyaddr_1:	buffer 1 physical address. Setting this to a value other than
   *		NULL enables double buffering mode.
   * @return:	0 on success or negative error code on failure.
   */
  static int ipu_init_channel_buffer(struct idmac_channel *ichan,
  				   enum pixel_fmt pixel_fmt,
  				   uint16_t width, uint16_t height,
  				   uint32_t stride,
  				   enum ipu_rotate_mode rot_mode,
  				   dma_addr_t phyaddr_0, dma_addr_t phyaddr_1)
  {
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
  	struct idmac *idmac = to_idmac(ichan->dma_chan.device);
  	struct ipu *ipu = to_ipu(idmac);
  	union chan_param_mem params = {};
  	unsigned long flags;
  	uint32_t reg;
  	uint32_t stride_bytes;
  
  	stride_bytes = stride * bytes_per_pixel(pixel_fmt);
  
  	if (stride_bytes % 4) {
  		dev_err(ipu->dev,
  			"Stride length must be 32-bit aligned, stride = %d, bytes = %d
  ",
  			stride, stride_bytes);
  		return -EINVAL;
  	}
  
  	/* IC channel's stride must be a multiple of 8 pixels */
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
686
  	if ((channel <= IDMAC_IC_13) && (stride % 8)) {
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
  		dev_err(ipu->dev, "Stride must be 8 pixel multiple
  ");
  		return -EINVAL;
  	}
  
  	/* Build parameter memory data for DMA channel */
  	ipu_ch_param_set_size(&params, pixel_fmt, width, height, stride_bytes);
  	ipu_ch_param_set_buffer(&params, phyaddr_0, phyaddr_1);
  	ipu_ch_param_set_rotation(&params, rot_mode);
  	/* Some channels (rotation) have restriction on burst length */
  	switch (channel) {
  	case IDMAC_IC_7:	/* Hangs with burst 8, 16, other values
  				   invalid - Table 44-30 */
  /*
  		ipu_ch_param_set_burst_size(&params, 8);
   */
  		break;
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  		/* In original code only IPU_PIX_FMT_RGB565 was setting burst */
  		ipu_ch_param_set_burst_size(&params, 16);
  		break;
  	case IDMAC_IC_0:
  	default:
  		break;
  	}
  
  	spin_lock_irqsave(&ipu->lock, flags);
  
  	ipu_write_param_mem(dma_param_addr(channel), (uint32_t *)&params, 10);
  
  	reg = idmac_read_ipureg(ipu, IPU_CHA_DB_MODE_SEL);
  
  	if (phyaddr_1)
  		reg |= 1UL << channel;
  	else
  		reg &= ~(1UL << channel);
  
  	idmac_write_ipureg(ipu, reg, IPU_CHA_DB_MODE_SEL);
  
  	ichan->status = IPU_CHANNEL_READY;
c74ef1f86   Luotao Fu   ipu_idmac: fix sp...
728
  	spin_unlock_irqrestore(&ipu->lock, flags);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
  
  	return 0;
  }
  
  /**
   * ipu_select_buffer() - mark a channel's buffer as ready.
   * @channel:	channel ID.
   * @buffer_n:	buffer number to mark ready.
   */
  static void ipu_select_buffer(enum ipu_channel channel, int buffer_n)
  {
  	/* No locking - this is a write-one-to-set register, cleared by IPU */
  	if (buffer_n == 0)
  		/* Mark buffer 0 as ready. */
  		idmac_write_ipureg(&ipu_data, 1UL << channel, IPU_CHA_BUF0_RDY);
  	else
  		/* Mark buffer 1 as ready. */
  		idmac_write_ipureg(&ipu_data, 1UL << channel, IPU_CHA_BUF1_RDY);
  }
  
  /**
   * ipu_update_channel_buffer() - update physical address of a channel buffer.
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
751
   * @ichan:	IDMAC channel.
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
752
753
754
   * @buffer_n:	buffer number to update.
   *		0 or 1 are the only valid values.
   * @phyaddr:	buffer physical address.
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
755
756
   */
  /* Called under spin_lock(_irqsave)(&ichan->lock) */
8f98781e0   Guennadi Liakhovetski   async-tx: fix buf...
757
758
  static void ipu_update_channel_buffer(struct idmac_channel *ichan,
  				      int buffer_n, dma_addr_t phyaddr)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
759
  {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
760
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
761
762
763
764
765
766
767
768
  	uint32_t reg;
  	unsigned long flags;
  
  	spin_lock_irqsave(&ipu_data.lock, flags);
  
  	if (buffer_n == 0) {
  		reg = idmac_read_ipureg(&ipu_data, IPU_CHA_BUF0_RDY);
  		if (reg & (1UL << channel)) {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
769
770
  			ipu_ic_disable_task(&ipu_data, channel);
  			ichan->status = IPU_CHANNEL_READY;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
771
772
773
774
775
776
777
778
779
  		}
  
  		/* 44.3.3.1.9 - Row Number 1 (WORD1, offset 0) */
  		idmac_write_ipureg(&ipu_data, dma_param_addr(channel) +
  				   0x0008UL, IPU_IMA_ADDR);
  		idmac_write_ipureg(&ipu_data, phyaddr, IPU_IMA_DATA);
  	} else {
  		reg = idmac_read_ipureg(&ipu_data, IPU_CHA_BUF1_RDY);
  		if (reg & (1UL << channel)) {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
780
781
  			ipu_ic_disable_task(&ipu_data, channel);
  			ichan->status = IPU_CHANNEL_READY;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
  		}
  
  		/* Check if double-buffering is already enabled */
  		reg = idmac_read_ipureg(&ipu_data, IPU_CHA_DB_MODE_SEL);
  
  		if (!(reg & (1UL << channel)))
  			idmac_write_ipureg(&ipu_data, reg | (1UL << channel),
  					   IPU_CHA_DB_MODE_SEL);
  
  		/* 44.3.3.1.9 - Row Number 1 (WORD1, offset 1) */
  		idmac_write_ipureg(&ipu_data, dma_param_addr(channel) +
  				   0x0009UL, IPU_IMA_ADDR);
  		idmac_write_ipureg(&ipu_data, phyaddr, IPU_IMA_DATA);
  	}
  
  	spin_unlock_irqrestore(&ipu_data.lock, flags);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
798
799
800
  }
  
  /* Called under spin_lock_irqsave(&ichan->lock) */
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
801
802
803
804
805
  static int ipu_submit_buffer(struct idmac_channel *ichan,
  	struct idmac_tx_desc *desc, struct scatterlist *sg, int buf_idx)
  {
  	unsigned int chan_id = ichan->dma_chan.chan_id;
  	struct device *dev = &ichan->dma_chan.dev->device;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
806
807
808
809
810
811
812
813
814
815
  
  	if (async_tx_test_ack(&desc->txd))
  		return -EINTR;
  
  	/*
  	 * On first invocation this shouldn't be necessary, the call to
  	 * ipu_init_channel_buffer() above will set addresses for us, so we
  	 * could make it conditional on status >= IPU_CHANNEL_ENABLED, but
  	 * doing it again shouldn't hurt either.
  	 */
8f98781e0   Guennadi Liakhovetski   async-tx: fix buf...
816
  	ipu_update_channel_buffer(ichan, buf_idx, sg_dma_address(sg));
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
817
818
819
820
821
822
823
824
825
826
  
  	ipu_select_buffer(chan_id, buf_idx);
  	dev_dbg(dev, "Updated sg %p on channel 0x%x buffer %d
  ",
  		sg, chan_id, buf_idx);
  
  	return 0;
  }
  
  /* Called under spin_lock_irqsave(&ichan->lock) */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
827
828
829
830
831
832
833
834
835
  static int ipu_submit_channel_buffers(struct idmac_channel *ichan,
  				      struct idmac_tx_desc *desc)
  {
  	struct scatterlist *sg;
  	int i, ret = 0;
  
  	for (i = 0, sg = desc->sg; i < 2 && sg; i++) {
  		if (!ichan->sg[i]) {
  			ichan->sg[i] = sg;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
836
  			ret = ipu_submit_buffer(ichan, desc, sg, i);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
837
838
  			if (ret < 0)
  				return ret;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
839
840
841
842
843
844
845
846
847
848
849
850
851
  			sg = sg_next(sg);
  		}
  	}
  
  	return ret;
  }
  
  static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx)
  {
  	struct idmac_tx_desc *desc = to_tx_desc(tx);
  	struct idmac_channel *ichan = to_idmac_chan(tx->chan);
  	struct idmac *idmac = to_idmac(tx->chan->device);
  	struct ipu *ipu = to_ipu(idmac);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
852
  	struct device *dev = &ichan->dma_chan.dev->device;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
853
854
  	dma_cookie_t cookie;
  	unsigned long flags;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
855
  	int ret;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
856
857
858
859
  
  	/* Sanity check */
  	if (!list_empty(&desc->list)) {
  		/* The descriptor doesn't belong to client */
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
860
861
  		dev_err(dev, "Descriptor %p not prepared!
  ", tx);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
862
863
864
865
  		return -EBUSY;
  	}
  
  	mutex_lock(&ichan->chan_mutex);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
866
  	async_tx_clear_ack(tx);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
  	if (ichan->status < IPU_CHANNEL_READY) {
  		struct idmac_video_param *video = &ichan->params.video;
  		/*
  		 * Initial buffer assignment - the first two sg-entries from
  		 * the descriptor will end up in the IDMAC buffers
  		 */
  		dma_addr_t dma_1 = sg_is_last(desc->sg) ? 0 :
  			sg_dma_address(&desc->sg[1]);
  
  		WARN_ON(ichan->sg[0] || ichan->sg[1]);
  
  		cookie = ipu_init_channel_buffer(ichan,
  						 video->out_pixel_fmt,
  						 video->out_width,
  						 video->out_height,
  						 video->out_stride,
  						 IPU_ROTATE_NONE,
  						 sg_dma_address(&desc->sg[0]),
  						 dma_1);
  		if (cookie < 0)
  			goto out;
  	}
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
889
890
  	dev_dbg(dev, "Submitting sg %p
  ", &desc->sg[0]);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
891
892
893
894
895
896
897
898
899
  
  	cookie = ichan->dma_chan.cookie;
  
  	if (++cookie < 0)
  		cookie = 1;
  
  	/* from dmaengine.h: "last cookie value returned to client" */
  	ichan->dma_chan.cookie = cookie;
  	tx->cookie = cookie;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
900
901
  
  	/* ipu->lock can be taken under ichan->lock, but not v.v. */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
902
  	spin_lock_irqsave(&ichan->lock, flags);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
903

5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
904
  	list_add_tail(&desc->list, &ichan->queue);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
905
906
  	/* submit_buffers() atomically verifies and fills empty sg slots */
  	ret = ipu_submit_channel_buffers(ichan, desc);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
907
  	spin_unlock_irqrestore(&ichan->lock, flags);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
908
909
910
911
  	if (ret < 0) {
  		cookie = ret;
  		goto dequeue;
  	}
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
912
  	if (ichan->status < IPU_CHANNEL_ENABLED) {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
913
  		ret = ipu_enable_channel(idmac, ichan);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
914
915
  		if (ret < 0) {
  			cookie = ret;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
916
  			goto dequeue;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
917
918
919
920
  		}
  	}
  
  	dump_idmac_reg(ipu);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
921
922
923
924
925
926
927
928
  dequeue:
  	if (cookie < 0) {
  		spin_lock_irqsave(&ichan->lock, flags);
  		list_del_init(&desc->list);
  		spin_unlock_irqrestore(&ichan->lock, flags);
  		tx->cookie = cookie;
  		ichan->dma_chan.cookie = cookie;
  	}
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
  out:
  	mutex_unlock(&ichan->chan_mutex);
  
  	return cookie;
  }
  
  /* Called with ichan->chan_mutex held */
  static int idmac_desc_alloc(struct idmac_channel *ichan, int n)
  {
  	struct idmac_tx_desc *desc = vmalloc(n * sizeof(struct idmac_tx_desc));
  	struct idmac *idmac = to_idmac(ichan->dma_chan.device);
  
  	if (!desc)
  		return -ENOMEM;
  
  	/* No interrupts, just disable the tasklet for a moment */
  	tasklet_disable(&to_ipu(idmac)->tasklet);
  
  	ichan->n_tx_desc = n;
  	ichan->desc = desc;
  	INIT_LIST_HEAD(&ichan->queue);
  	INIT_LIST_HEAD(&ichan->free_list);
  
  	while (n--) {
  		struct dma_async_tx_descriptor *txd = &desc->txd;
  
  		memset(txd, 0, sizeof(*txd));
  		dma_async_tx_descriptor_init(txd, &ichan->dma_chan);
  		txd->tx_submit		= idmac_tx_submit;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
  
  		list_add(&desc->list, &ichan->free_list);
  
  		desc++;
  	}
  
  	tasklet_enable(&to_ipu(idmac)->tasklet);
  
  	return 0;
  }
  
  /**
   * ipu_init_channel() - initialize an IPU channel.
   * @idmac:	IPU DMAC context.
   * @ichan:	pointer to the channel object.
   * @return      0 on success or negative error code on failure.
   */
  static int ipu_init_channel(struct idmac *idmac, struct idmac_channel *ichan)
  {
  	union ipu_channel_param *params = &ichan->params;
  	uint32_t ipu_conf;
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
  	unsigned long flags;
  	uint32_t reg;
  	struct ipu *ipu = to_ipu(idmac);
  	int ret = 0, n_desc = 0;
  
  	dev_dbg(ipu->dev, "init channel = %d
  ", channel);
  
  	if (channel != IDMAC_SDC_0 && channel != IDMAC_SDC_1 &&
  	    channel != IDMAC_IC_7)
  		return -EINVAL;
  
  	spin_lock_irqsave(&ipu->lock, flags);
  
  	switch (channel) {
  	case IDMAC_IC_7:
  		n_desc = 16;
  		reg = idmac_read_icreg(ipu, IC_CONF);
  		idmac_write_icreg(ipu, reg & ~IC_CONF_CSI_MEM_WR_EN, IC_CONF);
  		break;
  	case IDMAC_IC_0:
  		n_desc = 16;
  		reg = idmac_read_ipureg(ipu, IPU_FS_PROC_FLOW);
  		idmac_write_ipureg(ipu, reg & ~FS_ENC_IN_VALID, IPU_FS_PROC_FLOW);
  		ret = ipu_ic_init_prpenc(ipu, params, true);
  		break;
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  		n_desc = 4;
  	default:
  		break;
  	}
  
  	ipu->channel_init_mask |= 1L << channel;
  
  	/* Enable IPU sub module */
  	ipu_conf = idmac_read_ipureg(ipu, IPU_CONF) |
  		ipu_channel_conf_mask(channel);
  	idmac_write_ipureg(ipu, ipu_conf, IPU_CONF);
  
  	spin_unlock_irqrestore(&ipu->lock, flags);
  
  	if (n_desc && !ichan->desc)
  		ret = idmac_desc_alloc(ichan, n_desc);
  
  	dump_idmac_reg(ipu);
  
  	return ret;
  }
  
  /**
   * ipu_uninit_channel() - uninitialize an IPU channel.
   * @idmac:	IPU DMAC context.
   * @ichan:	pointer to the channel object.
   */
  static void ipu_uninit_channel(struct idmac *idmac, struct idmac_channel *ichan)
  {
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
  	unsigned long flags;
  	uint32_t reg;
  	unsigned long chan_mask = 1UL << channel;
  	uint32_t ipu_conf;
  	struct ipu *ipu = to_ipu(idmac);
  
  	spin_lock_irqsave(&ipu->lock, flags);
  
  	if (!(ipu->channel_init_mask & chan_mask)) {
  		dev_err(ipu->dev, "Channel already uninitialized %d
  ",
  			channel);
  		spin_unlock_irqrestore(&ipu->lock, flags);
  		return;
  	}
  
  	/* Reset the double buffer */
  	reg = idmac_read_ipureg(ipu, IPU_CHA_DB_MODE_SEL);
  	idmac_write_ipureg(ipu, reg & ~chan_mask, IPU_CHA_DB_MODE_SEL);
  
  	ichan->sec_chan_en = false;
  
  	switch (channel) {
  	case IDMAC_IC_7:
  		reg = idmac_read_icreg(ipu, IC_CONF);
  		idmac_write_icreg(ipu, reg & ~(IC_CONF_RWS_EN | IC_CONF_PRPENC_EN),
  			     IC_CONF);
  		break;
  	case IDMAC_IC_0:
  		reg = idmac_read_icreg(ipu, IC_CONF);
  		idmac_write_icreg(ipu, reg & ~(IC_CONF_PRPENC_EN | IC_CONF_PRPENC_CSC1),
  				  IC_CONF);
  		break;
  	case IDMAC_SDC_0:
  	case IDMAC_SDC_1:
  	default:
  		break;
  	}
  
  	ipu->channel_init_mask &= ~(1L << channel);
  
  	ipu_conf = idmac_read_ipureg(ipu, IPU_CONF) &
  		~ipu_channel_conf_mask(channel);
  	idmac_write_ipureg(ipu, ipu_conf, IPU_CONF);
  
  	spin_unlock_irqrestore(&ipu->lock, flags);
  
  	ichan->n_tx_desc = 0;
  	vfree(ichan->desc);
  	ichan->desc = NULL;
  }
  
  /**
   * ipu_disable_channel() - disable an IPU channel.
   * @idmac:		IPU DMAC context.
   * @ichan:		channel object pointer.
   * @wait_for_stop:	flag to set whether to wait for channel end of frame or
   *			return immediately.
   * @return:		0 on success or negative error code on failure.
   */
  static int ipu_disable_channel(struct idmac *idmac, struct idmac_channel *ichan,
  			       bool wait_for_stop)
  {
  	enum ipu_channel channel = ichan->dma_chan.chan_id;
  	struct ipu *ipu = to_ipu(idmac);
  	uint32_t reg;
  	unsigned long flags;
  	unsigned long chan_mask = 1UL << channel;
  	unsigned int timeout;
  
  	if (wait_for_stop && channel != IDMAC_SDC_1 && channel != IDMAC_SDC_0) {
  		timeout = 40;
  		/* This waiting always fails. Related to spurious irq problem */
  		while ((idmac_read_icreg(ipu, IDMAC_CHA_BUSY) & chan_mask) ||
  		       (ipu_channel_status(ipu, channel) == TASK_STAT_ACTIVE)) {
  			timeout--;
  			msleep(10);
  
  			if (!timeout) {
  				dev_dbg(ipu->dev,
  					"Warning: timeout waiting for channel %u to "
  					"stop: buf0_rdy = 0x%08X, buf1_rdy = 0x%08X, "
  					"busy = 0x%08X, tstat = 0x%08X
  ", channel,
  					idmac_read_ipureg(ipu, IPU_CHA_BUF0_RDY),
  					idmac_read_ipureg(ipu, IPU_CHA_BUF1_RDY),
  					idmac_read_icreg(ipu, IDMAC_CHA_BUSY),
  					idmac_read_ipureg(ipu, IPU_TASKS_STAT));
  				break;
  			}
  		}
  		dev_dbg(ipu->dev, "timeout = %d * 10ms
  ", 40 - timeout);
  	}
  	/* SDC BG and FG must be disabled before DMA is disabled */
  	if (wait_for_stop && (channel == IDMAC_SDC_0 ||
  			      channel == IDMAC_SDC_1)) {
  		for (timeout = 5;
  		     timeout && !ipu_irq_status(ichan->eof_irq); timeout--)
  			msleep(5);
  	}
  
  	spin_lock_irqsave(&ipu->lock, flags);
  
  	/* Disable IC task */
  	ipu_ic_disable_task(ipu, channel);
  
  	/* Disable DMA channel(s) */
  	reg = idmac_read_icreg(ipu, IDMAC_CHA_EN);
  	idmac_write_icreg(ipu, reg & ~chan_mask, IDMAC_CHA_EN);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1148
1149
1150
1151
  	spin_unlock_irqrestore(&ipu->lock, flags);
  
  	return 0;
  }
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
  static struct scatterlist *idmac_sg_next(struct idmac_channel *ichan,
  	struct idmac_tx_desc **desc, struct scatterlist *sg)
  {
  	struct scatterlist *sgnew = sg ? sg_next(sg) : NULL;
  
  	if (sgnew)
  		/* next sg-element in this list */
  		return sgnew;
  
  	if ((*desc)->list.next == &ichan->queue)
  		/* No more descriptors on the queue */
  		return NULL;
  
  	/* Fetch next descriptor */
  	*desc = list_entry((*desc)->list.next, struct idmac_tx_desc, list);
  	return (*desc)->sg;
  }
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
  /*
   * We have several possibilities here:
   * current BUF		next BUF
   *
   * not last sg		next not last sg
   * not last sg		next last sg
   * last sg		first sg from next descriptor
   * last sg		NULL
   *
   * Besides, the descriptor queue might be empty or not. We process all these
   * cases carefully.
   */
  static irqreturn_t idmac_interrupt(int irq, void *dev_id)
  {
  	struct idmac_channel *ichan = dev_id;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1184
  	struct device *dev = &ichan->dma_chan.dev->device;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1185
1186
1187
  	unsigned int chan_id = ichan->dma_chan.chan_id;
  	struct scatterlist **sg, *sgnext, *sgnew = NULL;
  	/* Next transfer descriptor */
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1188
  	struct idmac_tx_desc *desc, *descnew;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1189
1190
1191
  	dma_async_tx_callback callback;
  	void *callback_param;
  	bool done = false;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1192
1193
  	u32 ready0, ready1, curbuf, err;
  	unsigned long flags;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1194
1195
  
  	/* IDMAC has cleared the respective BUFx_RDY bit, we manage the buffer */
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
  	dev_dbg(dev, "IDMAC irq %d, buf %d
  ", irq, ichan->active_buffer);
  
  	spin_lock_irqsave(&ipu_data.lock, flags);
  
  	ready0	= idmac_read_ipureg(&ipu_data, IPU_CHA_BUF0_RDY);
  	ready1	= idmac_read_ipureg(&ipu_data, IPU_CHA_BUF1_RDY);
  	curbuf	= idmac_read_ipureg(&ipu_data, IPU_CHA_CUR_BUF);
  	err	= idmac_read_ipureg(&ipu_data, IPU_INT_STAT_4);
  
  	if (err & (1 << chan_id)) {
  		idmac_write_ipureg(&ipu_data, 1 << chan_id, IPU_INT_STAT_4);
  		spin_unlock_irqrestore(&ipu_data.lock, flags);
  		/*
  		 * Doing this
  		 * ichan->sg[0] = ichan->sg[1] = NULL;
  		 * you can force channel re-enable on the next tx_submit(), but
  		 * this is dirty - think about descriptors with multiple
  		 * sg elements.
  		 */
  		dev_warn(dev, "NFB4EOF on channel %d, ready %x, %x, cur %x
  ",
  			 chan_id, ready0, ready1, curbuf);
  		return IRQ_HANDLED;
  	}
  	spin_unlock_irqrestore(&ipu_data.lock, flags);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1222
1223
  	/* Other interrupts do not interfere with this channel */
  	spin_lock(&ichan->lock);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1224
1225
1226
1227
  	if (unlikely((ichan->active_buffer && (ready1 >> chan_id) & 1) ||
  		     (!ichan->active_buffer && (ready0 >> chan_id) & 1)
  		     )) {
  		spin_unlock(&ichan->lock);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1228
  		dev_dbg(dev,
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1229
1230
1231
1232
1233
1234
1235
1236
  			"IRQ with active buffer still ready on channel %x, "
  			"active %d, ready %x, %x!
  ", chan_id,
  			ichan->active_buffer, ready0, ready1);
  		return IRQ_NONE;
  	}
  
  	if (unlikely(list_empty(&ichan->queue))) {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1237
  		ichan->sg[ichan->active_buffer] = NULL;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1238
  		spin_unlock(&ichan->lock);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1239
  		dev_err(dev,
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
  			"IRQ without queued buffers on channel %x, active %d, "
  			"ready %x, %x!
  ", chan_id,
  			ichan->active_buffer, ready0, ready1);
  		return IRQ_NONE;
  	}
  
  	/*
  	 * active_buffer is a software flag, it shows which buffer we are
  	 * currently expecting back from the hardware, IDMAC should be
  	 * processing the other buffer already
  	 */
  	sg = &ichan->sg[ichan->active_buffer];
  	sgnext = ichan->sg[!ichan->active_buffer];
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
  	if (!*sg) {
  		spin_unlock(&ichan->lock);
  		return IRQ_HANDLED;
  	}
  
  	desc = list_entry(ichan->queue.next, struct idmac_tx_desc, list);
  	descnew = desc;
  
  	dev_dbg(dev, "IDMAC irq %d, dma 0x%08x, next dma 0x%08x, current %d, curbuf 0x%08x
  ",
  		irq, sg_dma_address(*sg), sgnext ? sg_dma_address(sgnext) : 0, ichan->active_buffer, curbuf);
  
  	/* Find the descriptor of sgnext */
  	sgnew = idmac_sg_next(ichan, &descnew, *sg);
  	if (sgnext != sgnew)
  		dev_err(dev, "Submitted buffer %p, next buffer %p
  ", sgnext, sgnew);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1271
1272
1273
1274
1275
  	/*
  	 * if sgnext == NULL sg must be the last element in a scatterlist and
  	 * queue must be empty
  	 */
  	if (unlikely(!sgnext)) {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1276
1277
1278
1279
1280
1281
1282
  		if (!WARN_ON(sg_next(*sg)))
  			dev_dbg(dev, "Underrun on channel %x
  ", chan_id);
  		ichan->sg[!ichan->active_buffer] = sgnew;
  
  		if (unlikely(sgnew)) {
  			ipu_submit_buffer(ichan, descnew, sgnew, !ichan->active_buffer);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1283
  		} else {
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1284
  			spin_lock_irqsave(&ipu_data.lock, flags);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1285
  			ipu_ic_disable_task(&ipu_data, chan_id);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1286
  			spin_unlock_irqrestore(&ipu_data.lock, flags);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1287
1288
1289
1290
  			ichan->status = IPU_CHANNEL_READY;
  			/* Continue to check for complete descriptor */
  		}
  	}
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1291
1292
  	/* Calculate and submit the next sg element */
  	sgnew = idmac_sg_next(ichan, &descnew, sgnew);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
  
  	if (unlikely(!sg_next(*sg)) || !sgnext) {
  		/*
  		 * Last element in scatterlist done, remove from the queue,
  		 * _init for debugging
  		 */
  		list_del_init(&desc->list);
  		done = true;
  	}
  
  	*sg = sgnew;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1304
1305
  	if (likely(sgnew) &&
  	    ipu_submit_buffer(ichan, descnew, sgnew, ichan->active_buffer) < 0) {
8f98781e0   Guennadi Liakhovetski   async-tx: fix buf...
1306
1307
  		callback = descnew->txd.callback;
  		callback_param = descnew->txd.callback_param;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1308
  		spin_unlock(&ichan->lock);
8f98781e0   Guennadi Liakhovetski   async-tx: fix buf...
1309
1310
  		if (callback)
  			callback(callback_param);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1311
  		spin_lock(&ichan->lock);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
  	}
  
  	/* Flip the active buffer - even if update above failed */
  	ichan->active_buffer = !ichan->active_buffer;
  	if (done)
  		ichan->completed = desc->txd.cookie;
  
  	callback = desc->txd.callback;
  	callback_param = desc->txd.callback_param;
  
  	spin_unlock(&ichan->lock);
  
  	if (done && (desc->txd.flags & DMA_PREP_INTERRUPT) && callback)
  		callback(callback_param);
  
  	return IRQ_HANDLED;
  }
  
  static void ipu_gc_tasklet(unsigned long arg)
  {
  	struct ipu *ipu = (struct ipu *)arg;
  	int i;
  
  	for (i = 0; i < IPU_CHANNELS_NUM; i++) {
  		struct idmac_channel *ichan = ipu->channel + i;
  		struct idmac_tx_desc *desc;
  		unsigned long flags;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1339
1340
  		struct scatterlist *sg;
  		int j, k;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1341
1342
1343
1344
1345
1346
  
  		for (j = 0; j < ichan->n_tx_desc; j++) {
  			desc = ichan->desc + j;
  			spin_lock_irqsave(&ichan->lock, flags);
  			if (async_tx_test_ack(&desc->txd)) {
  				list_move(&desc->list, &ichan->free_list);
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1347
1348
1349
1350
1351
1352
  				for_each_sg(desc->sg, sg, desc->sg_len, k) {
  					if (ichan->sg[0] == sg)
  						ichan->sg[0] = NULL;
  					else if (ichan->sg[1] == sg)
  						ichan->sg[1] = NULL;
  				}
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1353
1354
1355
1356
1357
1358
  				async_tx_clear_ack(&desc->txd);
  			}
  			spin_unlock_irqrestore(&ichan->lock, flags);
  		}
  	}
  }
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
1359
  /* Allocate and initialise a transfer descriptor. */
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
  static struct dma_async_tx_descriptor *idmac_prep_slave_sg(struct dma_chan *chan,
  		struct scatterlist *sgl, unsigned int sg_len,
  		enum dma_data_direction direction, unsigned long tx_flags)
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
  	struct idmac_tx_desc *desc = NULL;
  	struct dma_async_tx_descriptor *txd = NULL;
  	unsigned long flags;
  
  	/* We only can handle these three channels so far */
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1370
1371
  	if (chan->chan_id != IDMAC_SDC_0 && chan->chan_id != IDMAC_SDC_1 &&
  	    chan->chan_id != IDMAC_IC_7)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
  		return NULL;
  
  	if (direction != DMA_FROM_DEVICE && direction != DMA_TO_DEVICE) {
  		dev_err(chan->device->dev, "Invalid DMA direction %d!
  ", direction);
  		return NULL;
  	}
  
  	mutex_lock(&ichan->chan_mutex);
  
  	spin_lock_irqsave(&ichan->lock, flags);
  	if (!list_empty(&ichan->free_list)) {
  		desc = list_entry(ichan->free_list.next,
  				  struct idmac_tx_desc, list);
  
  		list_del_init(&desc->list);
  
  		desc->sg_len	= sg_len;
  		desc->sg	= sgl;
  		txd		= &desc->txd;
  		txd->flags	= tx_flags;
  	}
  	spin_unlock_irqrestore(&ichan->lock, flags);
  
  	mutex_unlock(&ichan->chan_mutex);
  
  	tasklet_schedule(&to_ipu(to_idmac(chan->device))->tasklet);
  
  	return txd;
  }
  
  /* Re-select the current buffer and re-activate the channel */
  static void idmac_issue_pending(struct dma_chan *chan)
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
  	struct idmac *idmac = to_idmac(chan->device);
  	struct ipu *ipu = to_ipu(idmac);
  	unsigned long flags;
  
  	/* This is not always needed, but doesn't hurt either */
  	spin_lock_irqsave(&ipu->lock, flags);
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1413
  	ipu_select_buffer(chan->chan_id, ichan->active_buffer);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1414
1415
1416
1417
1418
1419
1420
1421
1422
  	spin_unlock_irqrestore(&ipu->lock, flags);
  
  	/*
  	 * Might need to perform some parts of initialisation from
  	 * ipu_enable_channel(), but not all, we do not want to reset to buffer
  	 * 0, don't need to set priority again either, but re-enabling the task
  	 * and the channel might be a good idea.
  	 */
  }
058276303   Linus Walleij   DMAENGINE: extend...
1423
1424
  static int __idmac_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
  			   unsigned long arg)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1425
1426
1427
1428
1429
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
  	struct idmac *idmac = to_idmac(chan->device);
  	unsigned long flags;
  	int i;
c3635c78e   Linus Walleij   DMAENGINE: generi...
1430
1431
1432
  	/* Only supports DMA_TERMINATE_ALL */
  	if (cmd != DMA_TERMINATE_ALL)
  		return -ENXIO;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
  	ipu_disable_channel(idmac, ichan,
  			    ichan->status >= IPU_CHANNEL_ENABLED);
  
  	tasklet_disable(&to_ipu(idmac)->tasklet);
  
  	/* ichan->queue is modified in ISR, have to spinlock */
  	spin_lock_irqsave(&ichan->lock, flags);
  	list_splice_init(&ichan->queue, &ichan->free_list);
  
  	if (ichan->desc)
  		for (i = 0; i < ichan->n_tx_desc; i++) {
  			struct idmac_tx_desc *desc = ichan->desc + i;
  			if (list_empty(&desc->list))
  				/* Descriptor was prepared, but not submitted */
0149f7d5d   Guennadi Liakhovetski   dma: ipu_idmac dr...
1447
  				list_add(&desc->list, &ichan->free_list);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
  
  			async_tx_clear_ack(&desc->txd);
  		}
  
  	ichan->sg[0] = NULL;
  	ichan->sg[1] = NULL;
  	spin_unlock_irqrestore(&ichan->lock, flags);
  
  	tasklet_enable(&to_ipu(idmac)->tasklet);
  
  	ichan->status = IPU_CHANNEL_INITIALIZED;
c3635c78e   Linus Walleij   DMAENGINE: generi...
1459
1460
  
  	return 0;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1461
  }
058276303   Linus Walleij   DMAENGINE: extend...
1462
1463
  static int idmac_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
  			 unsigned long arg)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1464
1465
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
c3635c78e   Linus Walleij   DMAENGINE: generi...
1466
  	int ret;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1467
1468
  
  	mutex_lock(&ichan->chan_mutex);
058276303   Linus Walleij   DMAENGINE: extend...
1469
  	ret = __idmac_control(chan, cmd, arg);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1470
1471
  
  	mutex_unlock(&ichan->chan_mutex);
c3635c78e   Linus Walleij   DMAENGINE: generi...
1472
1473
  
  	return ret;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1474
  }
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1475
1476
1477
1478
1479
1480
1481
  #ifdef DEBUG
  static irqreturn_t ic_sof_irq(int irq, void *dev_id)
  {
  	struct idmac_channel *ichan = dev_id;
  	printk(KERN_DEBUG "Got SOF IRQ %d on Channel %d
  ",
  	       irq, ichan->dma_chan.chan_id);
ca50a51e8   Ben Nizette   ipu_idmac: Use di...
1482
  	disable_irq_nosync(irq);
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1483
1484
1485
1486
1487
1488
1489
1490
1491
  	return IRQ_HANDLED;
  }
  
  static irqreturn_t ic_eof_irq(int irq, void *dev_id)
  {
  	struct idmac_channel *ichan = dev_id;
  	printk(KERN_DEBUG "Got EOF IRQ %d on Channel %d
  ",
  	       irq, ichan->dma_chan.chan_id);
ca50a51e8   Ben Nizette   ipu_idmac: Use di...
1492
  	disable_irq_nosync(irq);
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1493
1494
1495
1496
1497
  	return IRQ_HANDLED;
  }
  
  static int ic_sof = -EINVAL, ic_eof = -EINVAL;
  #endif
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
  static int idmac_alloc_chan_resources(struct dma_chan *chan)
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
  	struct idmac *idmac = to_idmac(chan->device);
  	int ret;
  
  	/* dmaengine.c now guarantees to only offer free channels */
  	BUG_ON(chan->client_count > 1);
  	WARN_ON(ichan->status != IPU_CHANNEL_FREE);
  
  	chan->cookie		= 1;
  	ichan->completed	= -ENXIO;
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1510
  	ret = ipu_irq_map(chan->chan_id);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1511
1512
1513
1514
  	if (ret < 0)
  		goto eimap;
  
  	ichan->eof_irq = ret;
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1515
1516
1517
1518
1519
1520
  
  	/*
  	 * Important to first disable the channel, because maybe someone
  	 * used it before us, e.g., the bootloader
  	 */
  	ipu_disable_channel(idmac, ichan, true);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1521
1522
1523
1524
  
  	ret = ipu_init_channel(idmac, ichan);
  	if (ret < 0)
  		goto eichan;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1525
1526
1527
1528
  	ret = request_irq(ichan->eof_irq, idmac_interrupt, 0,
  			  ichan->eof_name, ichan);
  	if (ret < 0)
  		goto erirq;
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
  #ifdef DEBUG
  	if (chan->chan_id == IDMAC_IC_7) {
  		ic_sof = ipu_irq_map(69);
  		if (ic_sof > 0)
  			request_irq(ic_sof, ic_sof_irq, 0, "IC SOF", ichan);
  		ic_eof = ipu_irq_map(70);
  		if (ic_eof > 0)
  			request_irq(ic_eof, ic_eof_irq, 0, "IC EOF", ichan);
  	}
  #endif
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1539
1540
  
  	ichan->status = IPU_CHANNEL_INITIALIZED;
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1541
1542
1543
  	dev_dbg(&chan->dev->device, "Found channel 0x%x, irq %d
  ",
  		chan->chan_id, ichan->eof_irq);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1544
1545
  
  	return ret;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1546
  erirq:
8d47bae00   Guennadi Liakhovetski   dma: i.MX31 IPU D...
1547
1548
  	ipu_uninit_channel(idmac, ichan);
  eichan:
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1549
  	ipu_irq_unmap(chan->chan_id);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
  eimap:
  	return ret;
  }
  
  static void idmac_free_chan_resources(struct dma_chan *chan)
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
  	struct idmac *idmac = to_idmac(chan->device);
  
  	mutex_lock(&ichan->chan_mutex);
058276303   Linus Walleij   DMAENGINE: extend...
1560
  	__idmac_control(chan, DMA_TERMINATE_ALL, 0);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1561
1562
  
  	if (ichan->status > IPU_CHANNEL_FREE) {
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
  #ifdef DEBUG
  		if (chan->chan_id == IDMAC_IC_7) {
  			if (ic_sof > 0) {
  				free_irq(ic_sof, ichan);
  				ipu_irq_unmap(69);
  				ic_sof = -EINVAL;
  			}
  			if (ic_eof > 0) {
  				free_irq(ic_eof, ichan);
  				ipu_irq_unmap(70);
  				ic_eof = -EINVAL;
  			}
  		}
  #endif
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1577
  		free_irq(ichan->eof_irq, ichan);
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1578
  		ipu_irq_unmap(chan->chan_id);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
  	}
  
  	ichan->status = IPU_CHANNEL_FREE;
  
  	ipu_uninit_channel(idmac, ichan);
  
  	mutex_unlock(&ichan->chan_mutex);
  
  	tasklet_schedule(&to_ipu(idmac)->tasklet);
  }
079344818   Linus Walleij   DMAENGINE: generi...
1589
1590
  static enum dma_status idmac_tx_status(struct dma_chan *chan,
  		       dma_cookie_t cookie, struct dma_tx_state *txstate)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1591
1592
  {
  	struct idmac_channel *ichan = to_idmac_chan(chan);
bca346920   Dan Williams   dmaengine: provid...
1593
  	dma_set_tx_state(txstate, ichan->completed, chan->cookie, 0);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
  	if (cookie != chan->cookie)
  		return DMA_ERROR;
  	return DMA_SUCCESS;
  }
  
  static int __init ipu_idmac_init(struct ipu *ipu)
  {
  	struct idmac *idmac = &ipu->idmac;
  	struct dma_device *dma = &idmac->dma;
  	int i;
  
  	dma_cap_set(DMA_SLAVE, dma->cap_mask);
  	dma_cap_set(DMA_PRIVATE, dma->cap_mask);
  
  	/* Compulsory common fields */
  	dma->dev				= ipu->dev;
  	dma->device_alloc_chan_resources	= idmac_alloc_chan_resources;
  	dma->device_free_chan_resources		= idmac_free_chan_resources;
079344818   Linus Walleij   DMAENGINE: generi...
1612
  	dma->device_tx_status			= idmac_tx_status;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1613
1614
1615
1616
  	dma->device_issue_pending		= idmac_issue_pending;
  
  	/* Compulsory for DMA_SLAVE fields */
  	dma->device_prep_slave_sg		= idmac_prep_slave_sg;
c3635c78e   Linus Walleij   DMAENGINE: generi...
1617
  	dma->device_control			= idmac_control;
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
  
  	INIT_LIST_HEAD(&dma->channels);
  	for (i = 0; i < IPU_CHANNELS_NUM; i++) {
  		struct idmac_channel *ichan = ipu->channel + i;
  		struct dma_chan *dma_chan = &ichan->dma_chan;
  
  		spin_lock_init(&ichan->lock);
  		mutex_init(&ichan->chan_mutex);
  
  		ichan->status		= IPU_CHANNEL_FREE;
  		ichan->sec_chan_en	= false;
  		ichan->completed	= -ENXIO;
  		snprintf(ichan->eof_name, sizeof(ichan->eof_name), "IDMAC EOF %d", i);
  
  		dma_chan->device	= &idmac->dma;
  		dma_chan->cookie	= 1;
  		dma_chan->chan_id	= i;
8c6db1bbf   Guennadi Liakhovetski   dma: Add SoF and ...
1635
  		list_add_tail(&dma_chan->device_node, &dma->channels);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1636
1637
1638
1639
1640
1641
  	}
  
  	idmac_write_icreg(ipu, 0x00000070, IDMAC_CONF);
  
  	return dma_async_device_register(&idmac->dma);
  }
234f2df56   Guennadi Liakhovetski   dma: improve sect...
1642
  static void __exit ipu_idmac_exit(struct ipu *ipu)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1643
1644
1645
1646
1647
1648
  {
  	int i;
  	struct idmac *idmac = &ipu->idmac;
  
  	for (i = 0; i < IPU_CHANNELS_NUM; i++) {
  		struct idmac_channel *ichan = ipu->channel + i;
058276303   Linus Walleij   DMAENGINE: extend...
1649
  		idmac_control(&ichan->dma_chan, DMA_TERMINATE_ALL, 0);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1650
1651
1652
1653
1654
1655
1656
1657
1658
  		idmac_prep_slave_sg(&ichan->dma_chan, NULL, 0, DMA_NONE, 0);
  	}
  
  	dma_async_device_unregister(&idmac->dma);
  }
  
  /*****************************************************************************
   * IPU common probe / remove
   */
234f2df56   Guennadi Liakhovetski   dma: improve sect...
1659
  static int __init ipu_probe(struct platform_device *pdev)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
  {
  	struct ipu_platform_data *pdata = pdev->dev.platform_data;
  	struct resource *mem_ipu, *mem_ic;
  	int ret;
  
  	spin_lock_init(&ipu_data.lock);
  
  	mem_ipu	= platform_get_resource(pdev, IORESOURCE_MEM, 0);
  	mem_ic	= platform_get_resource(pdev, IORESOURCE_MEM, 1);
  	if (!pdata || !mem_ipu || !mem_ic)
  		return -EINVAL;
  
  	ipu_data.dev = &pdev->dev;
  
  	platform_set_drvdata(pdev, &ipu_data);
  
  	ret = platform_get_irq(pdev, 0);
  	if (ret < 0)
  		goto err_noirq;
  
  	ipu_data.irq_fn = ret;
  	ret = platform_get_irq(pdev, 1);
  	if (ret < 0)
  		goto err_noirq;
  
  	ipu_data.irq_err = ret;
  	ipu_data.irq_base = pdata->irq_base;
  
  	dev_dbg(&pdev->dev, "fn irq %u, err irq %u, irq-base %u
  ",
  		ipu_data.irq_fn, ipu_data.irq_err, ipu_data.irq_base);
  
  	/* Remap IPU common registers */
7dab35c0c   H Hartley Sweeten   dma: ipu_idmac.c:...
1693
  	ipu_data.reg_ipu = ioremap(mem_ipu->start, resource_size(mem_ipu));
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1694
1695
1696
1697
1698
1699
  	if (!ipu_data.reg_ipu) {
  		ret = -ENOMEM;
  		goto err_ioremap_ipu;
  	}
  
  	/* Remap Image Converter and Image DMA Controller registers */
7dab35c0c   H Hartley Sweeten   dma: ipu_idmac.c:...
1700
  	ipu_data.reg_ic = ioremap(mem_ic->start, resource_size(mem_ic));
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1701
1702
1703
1704
1705
1706
  	if (!ipu_data.reg_ic) {
  		ret = -ENOMEM;
  		goto err_ioremap_ic;
  	}
  
  	/* Get IPU clock */
9eb2eb8c4   Sascha Hauer   MX31 clkdev support
1707
  	ipu_data.ipu_clk = clk_get(&pdev->dev, NULL);
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
  	if (IS_ERR(ipu_data.ipu_clk)) {
  		ret = PTR_ERR(ipu_data.ipu_clk);
  		goto err_clk_get;
  	}
  
  	/* Make sure IPU HSP clock is running */
  	clk_enable(ipu_data.ipu_clk);
  
  	/* Disable all interrupts */
  	idmac_write_ipureg(&ipu_data, 0, IPU_INT_CTRL_1);
  	idmac_write_ipureg(&ipu_data, 0, IPU_INT_CTRL_2);
  	idmac_write_ipureg(&ipu_data, 0, IPU_INT_CTRL_3);
  	idmac_write_ipureg(&ipu_data, 0, IPU_INT_CTRL_4);
  	idmac_write_ipureg(&ipu_data, 0, IPU_INT_CTRL_5);
  
  	dev_dbg(&pdev->dev, "%s @ 0x%08lx, fn irq %u, err irq %u
  ", pdev->name,
  		(unsigned long)mem_ipu->start, ipu_data.irq_fn, ipu_data.irq_err);
  
  	ret = ipu_irq_attach_irq(&ipu_data, pdev);
  	if (ret < 0)
  		goto err_attach_irq;
  
  	/* Initialize DMA engine */
  	ret = ipu_idmac_init(&ipu_data);
  	if (ret < 0)
  		goto err_idmac_init;
  
  	tasklet_init(&ipu_data.tasklet, ipu_gc_tasklet, (unsigned long)&ipu_data);
  
  	ipu_data.dev = &pdev->dev;
  
  	dev_dbg(ipu_data.dev, "IPU initialized
  ");
  
  	return 0;
  
  err_idmac_init:
  err_attach_irq:
  	ipu_irq_detach_irq(&ipu_data, pdev);
  	clk_disable(ipu_data.ipu_clk);
  	clk_put(ipu_data.ipu_clk);
  err_clk_get:
  	iounmap(ipu_data.reg_ic);
  err_ioremap_ic:
  	iounmap(ipu_data.reg_ipu);
  err_ioremap_ipu:
  err_noirq:
  	dev_err(&pdev->dev, "Failed to probe IPU: %d
  ", ret);
  	return ret;
  }
234f2df56   Guennadi Liakhovetski   dma: improve sect...
1760
  static int __exit ipu_remove(struct platform_device *pdev)
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
  {
  	struct ipu *ipu = platform_get_drvdata(pdev);
  
  	ipu_idmac_exit(ipu);
  	ipu_irq_detach_irq(ipu, pdev);
  	clk_disable(ipu->ipu_clk);
  	clk_put(ipu->ipu_clk);
  	iounmap(ipu->reg_ic);
  	iounmap(ipu->reg_ipu);
  	tasklet_kill(&ipu->tasklet);
  	platform_set_drvdata(pdev, NULL);
  
  	return 0;
  }
  
  /*
   * We need two MEM resources - with IPU-common and Image Converter registers,
   * including PF_CONF and IDMAC_* registers, and two IRQs - function and error
   */
  static struct platform_driver ipu_platform_driver = {
  	.driver = {
  		.name	= "ipu-core",
  		.owner	= THIS_MODULE,
  	},
234f2df56   Guennadi Liakhovetski   dma: improve sect...
1785
  	.remove		= __exit_p(ipu_remove),
5296b56d1   Guennadi Liakhovetski   i.MX31: Image Pro...
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
  };
  
  static int __init ipu_init(void)
  {
  	return platform_driver_probe(&ipu_platform_driver, ipu_probe);
  }
  subsys_initcall(ipu_init);
  
  MODULE_DESCRIPTION("IPU core driver");
  MODULE_LICENSE("GPL v2");
  MODULE_AUTHOR("Guennadi Liakhovetski <lg@denx.de>");
  MODULE_ALIAS("platform:ipu-core");