1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2 // Copyright(c) 2015-17 Intel Corporation.
5 * Soundwire Intel Master Driver
8 #include <linux/acpi.h>
9 #include <linux/debugfs.h>
10 #include <linux/delay.h>
11 #include <linux/module.h>
12 #include <linux/interrupt.h>
14 #include <linux/platform_device.h>
15 #include <sound/pcm_params.h>
16 #include <linux/pm_runtime.h>
17 #include <sound/soc.h>
18 #include <linux/soundwire/sdw_registers.h>
19 #include <linux/soundwire/sdw.h>
20 #include <linux/soundwire/sdw_intel.h>
21 #include "cadence_master.h"
25 #define INTEL_MASTER_SUSPEND_DELAY_MS 3000
28 * debug/config flags for the Intel SoundWire Master.
30 * Since we may have multiple masters active, we can have up to 8
31 * flags reused in each byte, with master0 using the ls-byte, etc.
34 #define SDW_INTEL_MASTER_DISABLE_PM_RUNTIME BIT(0)
35 #define SDW_INTEL_MASTER_DISABLE_CLOCK_STOP BIT(1)
36 #define SDW_INTEL_MASTER_DISABLE_PM_RUNTIME_IDLE BIT(2)
37 #define SDW_INTEL_MASTER_DISABLE_MULTI_LINK BIT(3)
40 module_param_named(sdw_md_flags
, md_flags
, int, 0444);
41 MODULE_PARM_DESC(sdw_md_flags
, "SoundWire Intel Master device flags (0x0 all off)");
43 /* Intel SHIM Registers Definition */
44 #define SDW_SHIM_LCAP 0x0
45 #define SDW_SHIM_LCTL 0x4
46 #define SDW_SHIM_IPPTR 0x8
47 #define SDW_SHIM_SYNC 0xC
49 #define SDW_SHIM_CTLSCAP(x) (0x010 + 0x60 * (x))
50 #define SDW_SHIM_CTLS0CM(x) (0x012 + 0x60 * (x))
51 #define SDW_SHIM_CTLS1CM(x) (0x014 + 0x60 * (x))
52 #define SDW_SHIM_CTLS2CM(x) (0x016 + 0x60 * (x))
53 #define SDW_SHIM_CTLS3CM(x) (0x018 + 0x60 * (x))
54 #define SDW_SHIM_PCMSCAP(x) (0x020 + 0x60 * (x))
56 #define SDW_SHIM_PCMSYCHM(x, y) (0x022 + (0x60 * (x)) + (0x2 * (y)))
57 #define SDW_SHIM_PCMSYCHC(x, y) (0x042 + (0x60 * (x)) + (0x2 * (y)))
58 #define SDW_SHIM_PDMSCAP(x) (0x062 + 0x60 * (x))
59 #define SDW_SHIM_IOCTL(x) (0x06C + 0x60 * (x))
60 #define SDW_SHIM_CTMCTL(x) (0x06E + 0x60 * (x))
62 #define SDW_SHIM_WAKEEN 0x190
63 #define SDW_SHIM_WAKESTS 0x192
65 #define SDW_SHIM_LCTL_SPA BIT(0)
66 #define SDW_SHIM_LCTL_SPA_MASK GENMASK(3, 0)
67 #define SDW_SHIM_LCTL_CPA BIT(8)
68 #define SDW_SHIM_LCTL_CPA_MASK GENMASK(11, 8)
70 #define SDW_SHIM_SYNC_SYNCPRD_VAL_24 (24000 / SDW_CADENCE_GSYNC_KHZ - 1)
71 #define SDW_SHIM_SYNC_SYNCPRD_VAL_38_4 (38400 / SDW_CADENCE_GSYNC_KHZ - 1)
72 #define SDW_SHIM_SYNC_SYNCPRD GENMASK(14, 0)
73 #define SDW_SHIM_SYNC_SYNCCPU BIT(15)
74 #define SDW_SHIM_SYNC_CMDSYNC_MASK GENMASK(19, 16)
75 #define SDW_SHIM_SYNC_CMDSYNC BIT(16)
76 #define SDW_SHIM_SYNC_SYNCGO BIT(24)
78 #define SDW_SHIM_PCMSCAP_ISS GENMASK(3, 0)
79 #define SDW_SHIM_PCMSCAP_OSS GENMASK(7, 4)
80 #define SDW_SHIM_PCMSCAP_BSS GENMASK(12, 8)
82 #define SDW_SHIM_PCMSYCM_LCHN GENMASK(3, 0)
83 #define SDW_SHIM_PCMSYCM_HCHN GENMASK(7, 4)
84 #define SDW_SHIM_PCMSYCM_STREAM GENMASK(13, 8)
85 #define SDW_SHIM_PCMSYCM_DIR BIT(15)
87 #define SDW_SHIM_PDMSCAP_ISS GENMASK(3, 0)
88 #define SDW_SHIM_PDMSCAP_OSS GENMASK(7, 4)
89 #define SDW_SHIM_PDMSCAP_BSS GENMASK(12, 8)
90 #define SDW_SHIM_PDMSCAP_CPSS GENMASK(15, 13)
92 #define SDW_SHIM_IOCTL_MIF BIT(0)
93 #define SDW_SHIM_IOCTL_CO BIT(1)
94 #define SDW_SHIM_IOCTL_COE BIT(2)
95 #define SDW_SHIM_IOCTL_DO BIT(3)
96 #define SDW_SHIM_IOCTL_DOE BIT(4)
97 #define SDW_SHIM_IOCTL_BKE BIT(5)
98 #define SDW_SHIM_IOCTL_WPDD BIT(6)
99 #define SDW_SHIM_IOCTL_CIBD BIT(8)
100 #define SDW_SHIM_IOCTL_DIBD BIT(9)
102 #define SDW_SHIM_CTMCTL_DACTQE BIT(0)
103 #define SDW_SHIM_CTMCTL_DODS BIT(1)
104 #define SDW_SHIM_CTMCTL_DOAIS GENMASK(4, 3)
106 #define SDW_SHIM_WAKEEN_ENABLE BIT(0)
107 #define SDW_SHIM_WAKESTS_STATUS BIT(0)
109 /* Intel ALH Register definitions */
110 #define SDW_ALH_STRMZCFG(x) (0x000 + (0x4 * (x)))
111 #define SDW_ALH_NUM_STREAMS 64
113 #define SDW_ALH_STRMZCFG_DMAT_VAL 0x3
114 #define SDW_ALH_STRMZCFG_DMAT GENMASK(7, 0)
115 #define SDW_ALH_STRMZCFG_CHN GENMASK(19, 16)
117 enum intel_pdi_type
{
123 #define cdns_to_intel(_cdns) container_of(_cdns, struct sdw_intel, cdns)
126 * Read, write helpers for HW registers
128 static inline int intel_readl(void __iomem
*base
, int offset
)
130 return readl(base
+ offset
);
133 static inline void intel_writel(void __iomem
*base
, int offset
, int value
)
135 writel(value
, base
+ offset
);
138 static inline u16
intel_readw(void __iomem
*base
, int offset
)
140 return readw(base
+ offset
);
143 static inline void intel_writew(void __iomem
*base
, int offset
, u16 value
)
145 writew(value
, base
+ offset
);
148 static int intel_wait_bit(void __iomem
*base
, int offset
, u32 mask
, u32 target
)
154 reg_read
= readl(base
+ offset
);
155 if ((reg_read
& mask
) == target
)
159 usleep_range(50, 100);
160 } while (timeout
!= 0);
165 static int intel_clear_bit(void __iomem
*base
, int offset
, u32 value
, u32 mask
)
167 writel(value
, base
+ offset
);
168 return intel_wait_bit(base
, offset
, mask
, 0);
171 static int intel_set_bit(void __iomem
*base
, int offset
, u32 value
, u32 mask
)
173 writel(value
, base
+ offset
);
174 return intel_wait_bit(base
, offset
, mask
, mask
);
180 #ifdef CONFIG_DEBUG_FS
182 #define RD_BUF (2 * PAGE_SIZE)
184 static ssize_t
intel_sprintf(void __iomem
*mem
, bool l
,
185 char *buf
, size_t pos
, unsigned int reg
)
190 value
= intel_readl(mem
, reg
);
192 value
= intel_readw(mem
, reg
);
194 return scnprintf(buf
+ pos
, RD_BUF
- pos
, "%4x\t%4x\n", reg
, value
);
197 static int intel_reg_show(struct seq_file
*s_file
, void *data
)
199 struct sdw_intel
*sdw
= s_file
->private;
200 void __iomem
*s
= sdw
->link_res
->shim
;
201 void __iomem
*a
= sdw
->link_res
->alh
;
205 unsigned int links
, reg
;
207 buf
= kzalloc(RD_BUF
, GFP_KERNEL
);
211 links
= intel_readl(s
, SDW_SHIM_LCAP
) & GENMASK(2, 0);
213 ret
= scnprintf(buf
, RD_BUF
, "Register Value\n");
214 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\nShim\n");
216 for (i
= 0; i
< links
; i
++) {
217 reg
= SDW_SHIM_LCAP
+ i
* 4;
218 ret
+= intel_sprintf(s
, true, buf
, ret
, reg
);
221 for (i
= 0; i
< links
; i
++) {
222 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\nLink%d\n", i
);
223 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTLSCAP(i
));
224 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTLS0CM(i
));
225 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTLS1CM(i
));
226 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTLS2CM(i
));
227 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTLS3CM(i
));
228 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_PCMSCAP(i
));
230 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\n PCMSyCH registers\n");
233 * the value 10 is the number of PDIs. We will need a
234 * cleanup to remove hard-coded Intel configurations
235 * from cadence_master.c
237 for (j
= 0; j
< 10; j
++) {
238 ret
+= intel_sprintf(s
, false, buf
, ret
,
239 SDW_SHIM_PCMSYCHM(i
, j
));
240 ret
+= intel_sprintf(s
, false, buf
, ret
,
241 SDW_SHIM_PCMSYCHC(i
, j
));
243 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\n PDMSCAP, IOCTL, CTMCTL\n");
245 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_PDMSCAP(i
));
246 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_IOCTL(i
));
247 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_CTMCTL(i
));
250 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\nWake registers\n");
251 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_WAKEEN
);
252 ret
+= intel_sprintf(s
, false, buf
, ret
, SDW_SHIM_WAKESTS
);
254 ret
+= scnprintf(buf
+ ret
, RD_BUF
- ret
, "\nALH STRMzCFG\n");
255 for (i
= 0; i
< SDW_ALH_NUM_STREAMS
; i
++)
256 ret
+= intel_sprintf(a
, true, buf
, ret
, SDW_ALH_STRMZCFG(i
));
258 seq_printf(s_file
, "%s", buf
);
263 DEFINE_SHOW_ATTRIBUTE(intel_reg
);
265 static int intel_set_m_datamode(void *data
, u64 value
)
267 struct sdw_intel
*sdw
= data
;
268 struct sdw_bus
*bus
= &sdw
->cdns
.bus
;
270 if (value
> SDW_PORT_DATA_MODE_STATIC_1
)
273 /* Userspace changed the hardware state behind the kernel's back */
274 add_taint(TAINT_USER
, LOCKDEP_STILL_OK
);
276 bus
->params
.m_data_mode
= value
;
280 DEFINE_DEBUGFS_ATTRIBUTE(intel_set_m_datamode_fops
, NULL
,
281 intel_set_m_datamode
, "%llu\n");
283 static int intel_set_s_datamode(void *data
, u64 value
)
285 struct sdw_intel
*sdw
= data
;
286 struct sdw_bus
*bus
= &sdw
->cdns
.bus
;
288 if (value
> SDW_PORT_DATA_MODE_STATIC_1
)
291 /* Userspace changed the hardware state behind the kernel's back */
292 add_taint(TAINT_USER
, LOCKDEP_STILL_OK
);
294 bus
->params
.s_data_mode
= value
;
298 DEFINE_DEBUGFS_ATTRIBUTE(intel_set_s_datamode_fops
, NULL
,
299 intel_set_s_datamode
, "%llu\n");
301 static void intel_debugfs_init(struct sdw_intel
*sdw
)
303 struct dentry
*root
= sdw
->cdns
.bus
.debugfs
;
308 sdw
->debugfs
= debugfs_create_dir("intel-sdw", root
);
310 debugfs_create_file("intel-registers", 0400, sdw
->debugfs
, sdw
,
313 debugfs_create_file("intel-m-datamode", 0200, sdw
->debugfs
, sdw
,
314 &intel_set_m_datamode_fops
);
316 debugfs_create_file("intel-s-datamode", 0200, sdw
->debugfs
, sdw
,
317 &intel_set_s_datamode_fops
);
319 sdw_cdns_debugfs_init(&sdw
->cdns
, sdw
->debugfs
);
322 static void intel_debugfs_exit(struct sdw_intel
*sdw
)
324 debugfs_remove_recursive(sdw
->debugfs
);
327 static void intel_debugfs_init(struct sdw_intel
*sdw
) {}
328 static void intel_debugfs_exit(struct sdw_intel
*sdw
) {}
329 #endif /* CONFIG_DEBUG_FS */
335 static int intel_link_power_up(struct sdw_intel
*sdw
)
337 unsigned int link_id
= sdw
->instance
;
338 void __iomem
*shim
= sdw
->link_res
->shim
;
339 u32
*shim_mask
= sdw
->link_res
->shim_mask
;
340 struct sdw_bus
*bus
= &sdw
->cdns
.bus
;
341 struct sdw_master_prop
*prop
= &bus
->prop
;
342 u32 spa_mask
, cpa_mask
;
348 mutex_lock(sdw
->link_res
->shim_lock
);
351 * The hardware relies on an internal counter, typically 4kHz,
352 * to generate the SoundWire SSP - which defines a 'safe'
353 * synchronization point between commands and audio transport
354 * and allows for multi link synchronization. The SYNCPRD value
355 * is only dependent on the oscillator clock provided to
356 * the IP, so adjust based on _DSD properties reported in DSDT
357 * tables. The values reported are based on either 24MHz
358 * (CNL/CML) or 38.4 MHz (ICL/TGL+).
360 if (prop
->mclk_freq
% 6000000)
361 syncprd
= SDW_SHIM_SYNC_SYNCPRD_VAL_38_4
;
363 syncprd
= SDW_SHIM_SYNC_SYNCPRD_VAL_24
;
366 dev_dbg(sdw
->cdns
.dev
, "%s: powering up all links\n", __func__
);
368 /* we first need to program the SyncPRD/CPU registers */
369 dev_dbg(sdw
->cdns
.dev
,
370 "%s: first link up, programming SYNCPRD\n", __func__
);
372 /* set SyncPRD period */
373 sync_reg
= intel_readl(shim
, SDW_SHIM_SYNC
);
374 u32p_replace_bits(&sync_reg
, syncprd
, SDW_SHIM_SYNC_SYNCPRD
);
376 /* Set SyncCPU bit */
377 sync_reg
|= SDW_SHIM_SYNC_SYNCCPU
;
378 intel_writel(shim
, SDW_SHIM_SYNC
, sync_reg
);
380 /* Link power up sequence */
381 link_control
= intel_readl(shim
, SDW_SHIM_LCTL
);
383 /* only power-up enabled links */
384 spa_mask
= FIELD_PREP(SDW_SHIM_LCTL_SPA_MASK
, sdw
->link_res
->link_mask
);
385 cpa_mask
= FIELD_PREP(SDW_SHIM_LCTL_CPA_MASK
, sdw
->link_res
->link_mask
);
387 link_control
|= spa_mask
;
389 ret
= intel_set_bit(shim
, SDW_SHIM_LCTL
, link_control
, cpa_mask
);
391 dev_err(sdw
->cdns
.dev
, "Failed to power up link: %d\n", ret
);
395 /* SyncCPU will change once link is active */
396 ret
= intel_wait_bit(shim
, SDW_SHIM_SYNC
,
397 SDW_SHIM_SYNC_SYNCCPU
, 0);
399 dev_err(sdw
->cdns
.dev
,
400 "Failed to set SHIM_SYNC: %d\n", ret
);
405 *shim_mask
|= BIT(link_id
);
407 sdw
->cdns
.link_up
= true;
409 mutex_unlock(sdw
->link_res
->shim_lock
);
414 /* this needs to be called with shim_lock */
415 static void intel_shim_glue_to_master_ip(struct sdw_intel
*sdw
)
417 void __iomem
*shim
= sdw
->link_res
->shim
;
418 unsigned int link_id
= sdw
->instance
;
421 /* Switch to MIP from Glue logic */
422 ioctl
= intel_readw(shim
, SDW_SHIM_IOCTL(link_id
));
424 ioctl
&= ~(SDW_SHIM_IOCTL_DOE
);
425 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
426 usleep_range(10, 15);
428 ioctl
&= ~(SDW_SHIM_IOCTL_DO
);
429 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
430 usleep_range(10, 15);
432 ioctl
|= (SDW_SHIM_IOCTL_MIF
);
433 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
434 usleep_range(10, 15);
436 ioctl
&= ~(SDW_SHIM_IOCTL_BKE
);
437 ioctl
&= ~(SDW_SHIM_IOCTL_COE
);
438 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
439 usleep_range(10, 15);
441 /* at this point Master IP has full control of the I/Os */
444 /* this needs to be called with shim_lock */
445 static void intel_shim_master_ip_to_glue(struct sdw_intel
*sdw
)
447 unsigned int link_id
= sdw
->instance
;
448 void __iomem
*shim
= sdw
->link_res
->shim
;
452 ioctl
= intel_readw(shim
, SDW_SHIM_IOCTL(link_id
));
453 ioctl
|= SDW_SHIM_IOCTL_BKE
;
454 ioctl
|= SDW_SHIM_IOCTL_COE
;
455 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
456 usleep_range(10, 15);
458 ioctl
&= ~(SDW_SHIM_IOCTL_MIF
);
459 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
460 usleep_range(10, 15);
462 /* at this point Integration Glue has full control of the I/Os */
465 static int intel_shim_init(struct sdw_intel
*sdw
, bool clock_stop
)
467 void __iomem
*shim
= sdw
->link_res
->shim
;
468 unsigned int link_id
= sdw
->instance
;
470 u16 ioctl
= 0, act
= 0;
472 mutex_lock(sdw
->link_res
->shim_lock
);
474 /* Initialize Shim */
475 ioctl
|= SDW_SHIM_IOCTL_BKE
;
476 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
477 usleep_range(10, 15);
479 ioctl
|= SDW_SHIM_IOCTL_WPDD
;
480 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
481 usleep_range(10, 15);
483 ioctl
|= SDW_SHIM_IOCTL_DO
;
484 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
485 usleep_range(10, 15);
487 ioctl
|= SDW_SHIM_IOCTL_DOE
;
488 intel_writew(shim
, SDW_SHIM_IOCTL(link_id
), ioctl
);
489 usleep_range(10, 15);
491 intel_shim_glue_to_master_ip(sdw
);
493 u16p_replace_bits(&act
, 0x1, SDW_SHIM_CTMCTL_DOAIS
);
494 act
|= SDW_SHIM_CTMCTL_DACTQE
;
495 act
|= SDW_SHIM_CTMCTL_DODS
;
496 intel_writew(shim
, SDW_SHIM_CTMCTL(link_id
), act
);
497 usleep_range(10, 15);
499 mutex_unlock(sdw
->link_res
->shim_lock
);
504 static void intel_shim_wake(struct sdw_intel
*sdw
, bool wake_enable
)
506 void __iomem
*shim
= sdw
->link_res
->shim
;
507 unsigned int link_id
= sdw
->instance
;
508 u16 wake_en
, wake_sts
;
510 mutex_lock(sdw
->link_res
->shim_lock
);
511 wake_en
= intel_readw(shim
, SDW_SHIM_WAKEEN
);
514 /* Enable the wakeup */
515 wake_en
|= (SDW_SHIM_WAKEEN_ENABLE
<< link_id
);
516 intel_writew(shim
, SDW_SHIM_WAKEEN
, wake_en
);
518 /* Disable the wake up interrupt */
519 wake_en
&= ~(SDW_SHIM_WAKEEN_ENABLE
<< link_id
);
520 intel_writew(shim
, SDW_SHIM_WAKEEN
, wake_en
);
522 /* Clear wake status */
523 wake_sts
= intel_readw(shim
, SDW_SHIM_WAKESTS
);
524 wake_sts
|= (SDW_SHIM_WAKEEN_ENABLE
<< link_id
);
525 intel_writew(shim
, SDW_SHIM_WAKESTS_STATUS
, wake_sts
);
527 mutex_unlock(sdw
->link_res
->shim_lock
);
530 static int intel_link_power_down(struct sdw_intel
*sdw
)
532 u32 link_control
, spa_mask
, cpa_mask
;
533 unsigned int link_id
= sdw
->instance
;
534 void __iomem
*shim
= sdw
->link_res
->shim
;
535 u32
*shim_mask
= sdw
->link_res
->shim_mask
;
538 mutex_lock(sdw
->link_res
->shim_lock
);
540 intel_shim_master_ip_to_glue(sdw
);
542 if (!(*shim_mask
& BIT(link_id
)))
543 dev_err(sdw
->cdns
.dev
,
544 "%s: Unbalanced power-up/down calls\n", __func__
);
546 *shim_mask
&= ~BIT(link_id
);
550 dev_dbg(sdw
->cdns
.dev
, "%s: powering down all links\n", __func__
);
552 /* Link power down sequence */
553 link_control
= intel_readl(shim
, SDW_SHIM_LCTL
);
555 /* only power-down enabled links */
556 spa_mask
= FIELD_PREP(SDW_SHIM_LCTL_SPA_MASK
, ~sdw
->link_res
->link_mask
);
557 cpa_mask
= FIELD_PREP(SDW_SHIM_LCTL_CPA_MASK
, sdw
->link_res
->link_mask
);
559 link_control
&= spa_mask
;
561 ret
= intel_clear_bit(shim
, SDW_SHIM_LCTL
, link_control
, cpa_mask
);
564 link_control
= intel_readl(shim
, SDW_SHIM_LCTL
);
566 mutex_unlock(sdw
->link_res
->shim_lock
);
569 dev_err(sdw
->cdns
.dev
, "%s: could not power down link\n", __func__
);
574 sdw
->cdns
.link_up
= false;
578 static void intel_shim_sync_arm(struct sdw_intel
*sdw
)
580 void __iomem
*shim
= sdw
->link_res
->shim
;
583 mutex_lock(sdw
->link_res
->shim_lock
);
585 /* update SYNC register */
586 sync_reg
= intel_readl(shim
, SDW_SHIM_SYNC
);
587 sync_reg
|= (SDW_SHIM_SYNC_CMDSYNC
<< sdw
->instance
);
588 intel_writel(shim
, SDW_SHIM_SYNC
, sync_reg
);
590 mutex_unlock(sdw
->link_res
->shim_lock
);
593 static int intel_shim_sync_go_unlocked(struct sdw_intel
*sdw
)
595 void __iomem
*shim
= sdw
->link_res
->shim
;
599 /* Read SYNC register */
600 sync_reg
= intel_readl(shim
, SDW_SHIM_SYNC
);
603 * Set SyncGO bit to synchronously trigger a bank switch for
604 * all the masters. A write to SYNCGO bit clears CMDSYNC bit for all
607 sync_reg
|= SDW_SHIM_SYNC_SYNCGO
;
609 ret
= intel_clear_bit(shim
, SDW_SHIM_SYNC
, sync_reg
,
610 SDW_SHIM_SYNC_SYNCGO
);
613 dev_err(sdw
->cdns
.dev
, "SyncGO clear failed: %d\n", ret
);
618 static int intel_shim_sync_go(struct sdw_intel
*sdw
)
622 mutex_lock(sdw
->link_res
->shim_lock
);
624 ret
= intel_shim_sync_go_unlocked(sdw
);
626 mutex_unlock(sdw
->link_res
->shim_lock
);
634 static void intel_pdi_init(struct sdw_intel
*sdw
,
635 struct sdw_cdns_stream_config
*config
)
637 void __iomem
*shim
= sdw
->link_res
->shim
;
638 unsigned int link_id
= sdw
->instance
;
639 int pcm_cap
, pdm_cap
;
641 /* PCM Stream Capability */
642 pcm_cap
= intel_readw(shim
, SDW_SHIM_PCMSCAP(link_id
));
644 config
->pcm_bd
= FIELD_GET(SDW_SHIM_PCMSCAP_BSS
, pcm_cap
);
645 config
->pcm_in
= FIELD_GET(SDW_SHIM_PCMSCAP_ISS
, pcm_cap
);
646 config
->pcm_out
= FIELD_GET(SDW_SHIM_PCMSCAP_OSS
, pcm_cap
);
648 dev_dbg(sdw
->cdns
.dev
, "PCM cap bd:%d in:%d out:%d\n",
649 config
->pcm_bd
, config
->pcm_in
, config
->pcm_out
);
651 /* PDM Stream Capability */
652 pdm_cap
= intel_readw(shim
, SDW_SHIM_PDMSCAP(link_id
));
654 config
->pdm_bd
= FIELD_GET(SDW_SHIM_PDMSCAP_BSS
, pdm_cap
);
655 config
->pdm_in
= FIELD_GET(SDW_SHIM_PDMSCAP_ISS
, pdm_cap
);
656 config
->pdm_out
= FIELD_GET(SDW_SHIM_PDMSCAP_OSS
, pdm_cap
);
658 dev_dbg(sdw
->cdns
.dev
, "PDM cap bd:%d in:%d out:%d\n",
659 config
->pdm_bd
, config
->pdm_in
, config
->pdm_out
);
663 intel_pdi_get_ch_cap(struct sdw_intel
*sdw
, unsigned int pdi_num
, bool pcm
)
665 void __iomem
*shim
= sdw
->link_res
->shim
;
666 unsigned int link_id
= sdw
->instance
;
670 count
= intel_readw(shim
, SDW_SHIM_PCMSYCHC(link_id
, pdi_num
));
673 * WORKAROUND: on all existing Intel controllers, pdi
674 * number 2 reports channel count as 1 even though it
675 * supports 8 channels. Performing hardcoding for pdi
682 count
= intel_readw(shim
, SDW_SHIM_PDMSCAP(link_id
));
683 count
= FIELD_GET(SDW_SHIM_PDMSCAP_CPSS
, count
);
686 /* zero based values for channel count in register */
692 static int intel_pdi_get_ch_update(struct sdw_intel
*sdw
,
693 struct sdw_cdns_pdi
*pdi
,
694 unsigned int num_pdi
,
695 unsigned int *num_ch
, bool pcm
)
699 for (i
= 0; i
< num_pdi
; i
++) {
700 pdi
->ch_count
= intel_pdi_get_ch_cap(sdw
, pdi
->num
, pcm
);
701 ch_count
+= pdi
->ch_count
;
709 static int intel_pdi_stream_ch_update(struct sdw_intel
*sdw
,
710 struct sdw_cdns_streams
*stream
, bool pcm
)
712 intel_pdi_get_ch_update(sdw
, stream
->bd
, stream
->num_bd
,
713 &stream
->num_ch_bd
, pcm
);
715 intel_pdi_get_ch_update(sdw
, stream
->in
, stream
->num_in
,
716 &stream
->num_ch_in
, pcm
);
718 intel_pdi_get_ch_update(sdw
, stream
->out
, stream
->num_out
,
719 &stream
->num_ch_out
, pcm
);
724 static int intel_pdi_ch_update(struct sdw_intel
*sdw
)
726 /* First update PCM streams followed by PDM streams */
727 intel_pdi_stream_ch_update(sdw
, &sdw
->cdns
.pcm
, true);
728 intel_pdi_stream_ch_update(sdw
, &sdw
->cdns
.pdm
, false);
734 intel_pdi_shim_configure(struct sdw_intel
*sdw
, struct sdw_cdns_pdi
*pdi
)
736 void __iomem
*shim
= sdw
->link_res
->shim
;
737 unsigned int link_id
= sdw
->instance
;
740 /* the Bulk and PCM streams are not contiguous */
741 pdi
->intel_alh_id
= (link_id
* 16) + pdi
->num
+ 3;
743 pdi
->intel_alh_id
+= 2;
746 * Program stream parameters to stream SHIM register
747 * This is applicable for PCM stream only.
749 if (pdi
->type
!= SDW_STREAM_PCM
)
752 if (pdi
->dir
== SDW_DATA_DIR_RX
)
753 pdi_conf
|= SDW_SHIM_PCMSYCM_DIR
;
755 pdi_conf
&= ~(SDW_SHIM_PCMSYCM_DIR
);
757 u32p_replace_bits(&pdi_conf
, pdi
->intel_alh_id
, SDW_SHIM_PCMSYCM_STREAM
);
758 u32p_replace_bits(&pdi_conf
, pdi
->l_ch_num
, SDW_SHIM_PCMSYCM_LCHN
);
759 u32p_replace_bits(&pdi_conf
, pdi
->h_ch_num
, SDW_SHIM_PCMSYCM_HCHN
);
761 intel_writew(shim
, SDW_SHIM_PCMSYCHM(link_id
, pdi
->num
), pdi_conf
);
765 intel_pdi_alh_configure(struct sdw_intel
*sdw
, struct sdw_cdns_pdi
*pdi
)
767 void __iomem
*alh
= sdw
->link_res
->alh
;
768 unsigned int link_id
= sdw
->instance
;
771 /* the Bulk and PCM streams are not contiguous */
772 pdi
->intel_alh_id
= (link_id
* 16) + pdi
->num
+ 3;
774 pdi
->intel_alh_id
+= 2;
776 /* Program Stream config ALH register */
777 conf
= intel_readl(alh
, SDW_ALH_STRMZCFG(pdi
->intel_alh_id
));
779 u32p_replace_bits(&conf
, SDW_ALH_STRMZCFG_DMAT_VAL
, SDW_ALH_STRMZCFG_DMAT
);
780 u32p_replace_bits(&conf
, pdi
->ch_count
- 1, SDW_ALH_STRMZCFG_CHN
);
782 intel_writel(alh
, SDW_ALH_STRMZCFG(pdi
->intel_alh_id
), conf
);
785 static int intel_params_stream(struct sdw_intel
*sdw
,
786 struct snd_pcm_substream
*substream
,
787 struct snd_soc_dai
*dai
,
788 struct snd_pcm_hw_params
*hw_params
,
789 int link_id
, int alh_stream_id
)
791 struct sdw_intel_link_res
*res
= sdw
->link_res
;
792 struct sdw_intel_stream_params_data params_data
;
794 params_data
.substream
= substream
;
795 params_data
.dai
= dai
;
796 params_data
.hw_params
= hw_params
;
797 params_data
.link_id
= link_id
;
798 params_data
.alh_stream_id
= alh_stream_id
;
800 if (res
->ops
&& res
->ops
->params_stream
&& res
->dev
)
801 return res
->ops
->params_stream(res
->dev
,
806 static int intel_free_stream(struct sdw_intel
*sdw
,
807 struct snd_pcm_substream
*substream
,
808 struct snd_soc_dai
*dai
,
811 struct sdw_intel_link_res
*res
= sdw
->link_res
;
812 struct sdw_intel_stream_free_data free_data
;
814 free_data
.substream
= substream
;
816 free_data
.link_id
= link_id
;
818 if (res
->ops
&& res
->ops
->free_stream
&& res
->dev
)
819 return res
->ops
->free_stream(res
->dev
,
826 * bank switch routines
829 static int intel_pre_bank_switch(struct sdw_bus
*bus
)
831 struct sdw_cdns
*cdns
= bus_to_cdns(bus
);
832 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
834 /* Write to register only for multi-link */
835 if (!bus
->multi_link
)
838 intel_shim_sync_arm(sdw
);
843 static int intel_post_bank_switch(struct sdw_bus
*bus
)
845 struct sdw_cdns
*cdns
= bus_to_cdns(bus
);
846 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
847 void __iomem
*shim
= sdw
->link_res
->shim
;
850 /* Write to register only for multi-link */
851 if (!bus
->multi_link
)
854 mutex_lock(sdw
->link_res
->shim_lock
);
856 /* Read SYNC register */
857 sync_reg
= intel_readl(shim
, SDW_SHIM_SYNC
);
860 * post_bank_switch() ops is called from the bus in loop for
861 * all the Masters in the steam with the expectation that
862 * we trigger the bankswitch for the only first Master in the list
863 * and do nothing for the other Masters
865 * So, set the SYNCGO bit only if CMDSYNC bit is set for any Master.
867 if (!(sync_reg
& SDW_SHIM_SYNC_CMDSYNC_MASK
)) {
872 ret
= intel_shim_sync_go_unlocked(sdw
);
874 mutex_unlock(sdw
->link_res
->shim_lock
);
877 dev_err(sdw
->cdns
.dev
, "Post bank switch failed: %d\n", ret
);
886 static int intel_startup(struct snd_pcm_substream
*substream
,
887 struct snd_soc_dai
*dai
)
889 struct sdw_cdns
*cdns
= snd_soc_dai_get_drvdata(dai
);
892 ret
= pm_runtime_get_sync(cdns
->dev
);
893 if (ret
< 0 && ret
!= -EACCES
) {
894 dev_err_ratelimited(cdns
->dev
,
895 "pm_runtime_get_sync failed in %s, ret %d\n",
897 pm_runtime_put_noidle(cdns
->dev
);
903 static int intel_hw_params(struct snd_pcm_substream
*substream
,
904 struct snd_pcm_hw_params
*params
,
905 struct snd_soc_dai
*dai
)
907 struct sdw_cdns
*cdns
= snd_soc_dai_get_drvdata(dai
);
908 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
909 struct sdw_cdns_dma_data
*dma
;
910 struct sdw_cdns_pdi
*pdi
;
911 struct sdw_stream_config sconfig
;
912 struct sdw_port_config
*pconfig
;
917 dma
= snd_soc_dai_get_dma_data(dai
, substream
);
921 ch
= params_channels(params
);
922 if (substream
->stream
== SNDRV_PCM_STREAM_CAPTURE
)
923 dir
= SDW_DATA_DIR_RX
;
925 dir
= SDW_DATA_DIR_TX
;
927 if (dma
->stream_type
== SDW_STREAM_PDM
)
931 pdi
= sdw_cdns_alloc_pdi(cdns
, &cdns
->pcm
, ch
, dir
, dai
->id
);
933 pdi
= sdw_cdns_alloc_pdi(cdns
, &cdns
->pdm
, ch
, dir
, dai
->id
);
940 /* do run-time configurations for SHIM, ALH and PDI/PORT */
941 intel_pdi_shim_configure(sdw
, pdi
);
942 intel_pdi_alh_configure(sdw
, pdi
);
943 sdw_cdns_config_stream(cdns
, ch
, dir
, pdi
);
945 /* store pdi and hw_params, may be needed in prepare step */
946 dma
->suspended
= false;
948 dma
->hw_params
= params
;
950 /* Inform DSP about PDI stream number */
951 ret
= intel_params_stream(sdw
, substream
, dai
, params
,
957 sconfig
.direction
= dir
;
958 sconfig
.ch_count
= ch
;
959 sconfig
.frame_rate
= params_rate(params
);
960 sconfig
.type
= dma
->stream_type
;
962 if (dma
->stream_type
== SDW_STREAM_PDM
) {
963 sconfig
.frame_rate
*= 50;
966 sconfig
.bps
= snd_pcm_format_width(params_format(params
));
969 /* Port configuration */
970 pconfig
= kcalloc(1, sizeof(*pconfig
), GFP_KERNEL
);
976 pconfig
->num
= pdi
->num
;
977 pconfig
->ch_mask
= (1 << ch
) - 1;
979 ret
= sdw_stream_add_master(&cdns
->bus
, &sconfig
,
980 pconfig
, 1, dma
->stream
);
982 dev_err(cdns
->dev
, "add master to stream failed:%d\n", ret
);
989 static int intel_prepare(struct snd_pcm_substream
*substream
,
990 struct snd_soc_dai
*dai
)
992 struct sdw_cdns
*cdns
= snd_soc_dai_get_drvdata(dai
);
993 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
994 struct sdw_cdns_dma_data
*dma
;
998 dma
= snd_soc_dai_get_dma_data(dai
, substream
);
1000 dev_err(dai
->dev
, "failed to get dma data in %s",
1005 if (dma
->suspended
) {
1006 dma
->suspended
= false;
1009 * .prepare() is called after system resume, where we
1010 * need to reinitialize the SHIM/ALH/Cadence IP.
1011 * .prepare() is also called to deal with underflows,
1012 * but in those cases we cannot touch ALH/SHIM
1016 /* configure stream */
1017 ch
= params_channels(dma
->hw_params
);
1018 if (substream
->stream
== SNDRV_PCM_STREAM_CAPTURE
)
1019 dir
= SDW_DATA_DIR_RX
;
1021 dir
= SDW_DATA_DIR_TX
;
1023 intel_pdi_shim_configure(sdw
, dma
->pdi
);
1024 intel_pdi_alh_configure(sdw
, dma
->pdi
);
1025 sdw_cdns_config_stream(cdns
, ch
, dir
, dma
->pdi
);
1027 /* Inform DSP about PDI stream number */
1028 ret
= intel_params_stream(sdw
, substream
, dai
,
1031 dma
->pdi
->intel_alh_id
);
1038 intel_hw_free(struct snd_pcm_substream
*substream
, struct snd_soc_dai
*dai
)
1040 struct sdw_cdns
*cdns
= snd_soc_dai_get_drvdata(dai
);
1041 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1042 struct sdw_cdns_dma_data
*dma
;
1045 dma
= snd_soc_dai_get_dma_data(dai
, substream
);
1050 * The sdw stream state will transition to RELEASED when stream->
1051 * master_list is empty. So the stream state will transition to
1052 * DEPREPARED for the first cpu-dai and to RELEASED for the last
1055 ret
= sdw_stream_remove_master(&cdns
->bus
, dma
->stream
);
1057 dev_err(dai
->dev
, "remove master from stream %s failed: %d\n",
1058 dma
->stream
->name
, ret
);
1062 ret
= intel_free_stream(sdw
, substream
, dai
, sdw
->instance
);
1064 dev_err(dai
->dev
, "intel_free_stream: failed %d", ret
);
1068 dma
->hw_params
= NULL
;
1074 static void intel_shutdown(struct snd_pcm_substream
*substream
,
1075 struct snd_soc_dai
*dai
)
1077 struct sdw_cdns
*cdns
= snd_soc_dai_get_drvdata(dai
);
1079 pm_runtime_mark_last_busy(cdns
->dev
);
1080 pm_runtime_put_autosuspend(cdns
->dev
);
1083 static int intel_component_dais_suspend(struct snd_soc_component
*component
)
1085 struct sdw_cdns_dma_data
*dma
;
1086 struct snd_soc_dai
*dai
;
1088 for_each_component_dais(component
, dai
) {
1090 * we don't have a .suspend dai_ops, and we don't have access
1091 * to the substream, so let's mark both capture and playback
1092 * DMA contexts as suspended
1094 dma
= dai
->playback_dma_data
;
1096 dma
->suspended
= true;
1098 dma
= dai
->capture_dma_data
;
1100 dma
->suspended
= true;
1106 static int intel_pcm_set_sdw_stream(struct snd_soc_dai
*dai
,
1107 void *stream
, int direction
)
1109 return cdns_set_sdw_stream(dai
, stream
, true, direction
);
1112 static int intel_pdm_set_sdw_stream(struct snd_soc_dai
*dai
,
1113 void *stream
, int direction
)
1115 return cdns_set_sdw_stream(dai
, stream
, false, direction
);
1118 static void *intel_get_sdw_stream(struct snd_soc_dai
*dai
,
1121 struct sdw_cdns_dma_data
*dma
;
1123 if (direction
== SNDRV_PCM_STREAM_PLAYBACK
)
1124 dma
= dai
->playback_dma_data
;
1126 dma
= dai
->capture_dma_data
;
1129 return ERR_PTR(-EINVAL
);
1134 static const struct snd_soc_dai_ops intel_pcm_dai_ops
= {
1135 .startup
= intel_startup
,
1136 .hw_params
= intel_hw_params
,
1137 .prepare
= intel_prepare
,
1138 .hw_free
= intel_hw_free
,
1139 .shutdown
= intel_shutdown
,
1140 .set_sdw_stream
= intel_pcm_set_sdw_stream
,
1141 .get_sdw_stream
= intel_get_sdw_stream
,
1144 static const struct snd_soc_dai_ops intel_pdm_dai_ops
= {
1145 .startup
= intel_startup
,
1146 .hw_params
= intel_hw_params
,
1147 .prepare
= intel_prepare
,
1148 .hw_free
= intel_hw_free
,
1149 .shutdown
= intel_shutdown
,
1150 .set_sdw_stream
= intel_pdm_set_sdw_stream
,
1151 .get_sdw_stream
= intel_get_sdw_stream
,
1154 static const struct snd_soc_component_driver dai_component
= {
1155 .name
= "soundwire",
1156 .suspend
= intel_component_dais_suspend
1159 static int intel_create_dai(struct sdw_cdns
*cdns
,
1160 struct snd_soc_dai_driver
*dais
,
1161 enum intel_pdi_type type
,
1162 u32 num
, u32 off
, u32 max_ch
, bool pcm
)
1169 /* TODO: Read supported rates/formats from hardware */
1170 for (i
= off
; i
< (off
+ num
); i
++) {
1171 dais
[i
].name
= devm_kasprintf(cdns
->dev
, GFP_KERNEL
,
1177 if (type
== INTEL_PDI_BD
|| type
== INTEL_PDI_OUT
) {
1178 dais
[i
].playback
.channels_min
= 1;
1179 dais
[i
].playback
.channels_max
= max_ch
;
1180 dais
[i
].playback
.rates
= SNDRV_PCM_RATE_48000
;
1181 dais
[i
].playback
.formats
= SNDRV_PCM_FMTBIT_S16_LE
;
1184 if (type
== INTEL_PDI_BD
|| type
== INTEL_PDI_IN
) {
1185 dais
[i
].capture
.channels_min
= 1;
1186 dais
[i
].capture
.channels_max
= max_ch
;
1187 dais
[i
].capture
.rates
= SNDRV_PCM_RATE_48000
;
1188 dais
[i
].capture
.formats
= SNDRV_PCM_FMTBIT_S16_LE
;
1192 dais
[i
].ops
= &intel_pcm_dai_ops
;
1194 dais
[i
].ops
= &intel_pdm_dai_ops
;
1200 static int intel_register_dai(struct sdw_intel
*sdw
)
1202 struct sdw_cdns
*cdns
= &sdw
->cdns
;
1203 struct sdw_cdns_streams
*stream
;
1204 struct snd_soc_dai_driver
*dais
;
1205 int num_dai
, ret
, off
= 0;
1207 /* DAIs are created based on total number of PDIs supported */
1208 num_dai
= cdns
->pcm
.num_pdi
+ cdns
->pdm
.num_pdi
;
1210 dais
= devm_kcalloc(cdns
->dev
, num_dai
, sizeof(*dais
), GFP_KERNEL
);
1214 /* Create PCM DAIs */
1215 stream
= &cdns
->pcm
;
1217 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_IN
, cdns
->pcm
.num_in
,
1218 off
, stream
->num_ch_in
, true);
1222 off
+= cdns
->pcm
.num_in
;
1223 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_OUT
, cdns
->pcm
.num_out
,
1224 off
, stream
->num_ch_out
, true);
1228 off
+= cdns
->pcm
.num_out
;
1229 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_BD
, cdns
->pcm
.num_bd
,
1230 off
, stream
->num_ch_bd
, true);
1234 /* Create PDM DAIs */
1235 stream
= &cdns
->pdm
;
1236 off
+= cdns
->pcm
.num_bd
;
1237 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_IN
, cdns
->pdm
.num_in
,
1238 off
, stream
->num_ch_in
, false);
1242 off
+= cdns
->pdm
.num_in
;
1243 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_OUT
, cdns
->pdm
.num_out
,
1244 off
, stream
->num_ch_out
, false);
1248 off
+= cdns
->pdm
.num_out
;
1249 ret
= intel_create_dai(cdns
, dais
, INTEL_PDI_BD
, cdns
->pdm
.num_bd
,
1250 off
, stream
->num_ch_bd
, false);
1254 return snd_soc_register_component(cdns
->dev
, &dai_component
,
1258 static int sdw_master_read_intel_prop(struct sdw_bus
*bus
)
1260 struct sdw_master_prop
*prop
= &bus
->prop
;
1261 struct fwnode_handle
*link
;
1265 /* Find master handle */
1266 snprintf(name
, sizeof(name
),
1267 "mipi-sdw-link-%d-subproperties", bus
->link_id
);
1269 link
= device_get_named_child_node(bus
->dev
, name
);
1271 dev_err(bus
->dev
, "Master node %s not found\n", name
);
1275 fwnode_property_read_u32(link
,
1276 "intel-sdw-ip-clock",
1279 /* the values reported by BIOS are the 2x clock, not the bus clock */
1280 prop
->mclk_freq
/= 2;
1282 fwnode_property_read_u32(link
,
1286 if (quirk_mask
& SDW_INTEL_QUIRK_MASK_BUS_DISABLE
)
1287 prop
->hw_disabled
= true;
1292 static int intel_prop_read(struct sdw_bus
*bus
)
1294 /* Initialize with default handler to read all DisCo properties */
1295 sdw_master_read_prop(bus
);
1297 /* read Intel-specific properties */
1298 sdw_master_read_intel_prop(bus
);
1303 static struct sdw_master_ops sdw_intel_ops
= {
1304 .read_prop
= sdw_master_read_prop
,
1305 .xfer_msg
= cdns_xfer_msg
,
1306 .xfer_msg_defer
= cdns_xfer_msg_defer
,
1307 .reset_page_addr
= cdns_reset_page_addr
,
1308 .set_bus_conf
= cdns_bus_conf
,
1309 .pre_bank_switch
= intel_pre_bank_switch
,
1310 .post_bank_switch
= intel_post_bank_switch
,
1313 static int intel_init(struct sdw_intel
*sdw
)
1317 /* Initialize shim and controller */
1318 intel_link_power_up(sdw
);
1320 clock_stop
= sdw_cdns_is_clock_stop(&sdw
->cdns
);
1322 intel_shim_init(sdw
, clock_stop
);
1330 static int intel_master_probe(struct platform_device
*pdev
)
1332 struct device
*dev
= &pdev
->dev
;
1333 struct sdw_intel
*sdw
;
1334 struct sdw_cdns
*cdns
;
1335 struct sdw_bus
*bus
;
1338 sdw
= devm_kzalloc(dev
, sizeof(*sdw
), GFP_KERNEL
);
1345 sdw
->instance
= pdev
->id
;
1346 sdw
->link_res
= dev_get_platdata(dev
);
1348 cdns
->registers
= sdw
->link_res
->registers
;
1349 cdns
->instance
= sdw
->instance
;
1350 cdns
->msg_count
= 0;
1352 bus
->link_id
= pdev
->id
;
1354 sdw_cdns_probe(cdns
);
1356 /* Set property read ops */
1357 sdw_intel_ops
.read_prop
= intel_prop_read
;
1358 bus
->ops
= &sdw_intel_ops
;
1360 /* set driver data, accessed by snd_soc_dai_get_drvdata() */
1361 dev_set_drvdata(dev
, cdns
);
1363 /* use generic bandwidth allocation algorithm */
1364 sdw
->cdns
.bus
.compute_params
= sdw_compute_params
;
1366 ret
= sdw_bus_master_add(bus
, dev
, dev
->fwnode
);
1368 dev_err(dev
, "sdw_bus_master_add fail: %d\n", ret
);
1372 if (bus
->prop
.hw_disabled
)
1374 "SoundWire master %d is disabled, will be ignored\n",
1377 * Ignore BIOS err_threshold, it's a really bad idea when dealing
1378 * with multiple hardware synchronized links
1380 bus
->prop
.err_threshold
= 0;
1385 int intel_master_startup(struct platform_device
*pdev
)
1387 struct sdw_cdns_stream_config config
;
1388 struct device
*dev
= &pdev
->dev
;
1389 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1390 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1391 struct sdw_bus
*bus
= &cdns
->bus
;
1394 u32 clock_stop_quirks
;
1397 if (bus
->prop
.hw_disabled
) {
1399 "SoundWire master %d is disabled, ignoring\n",
1404 link_flags
= md_flags
>> (bus
->link_id
* 8);
1405 multi_link
= !(link_flags
& SDW_INTEL_MASTER_DISABLE_MULTI_LINK
);
1407 dev_dbg(dev
, "Multi-link is disabled\n");
1408 bus
->multi_link
= false;
1411 * hardware-based synchronization is required regardless
1412 * of the number of segments used by a stream: SSP-based
1413 * synchronization is gated by gsync when the multi-master
1416 bus
->multi_link
= true;
1417 bus
->hw_sync_min_links
= 1;
1420 /* Initialize shim, controller */
1421 ret
= intel_init(sdw
);
1425 /* Read the PDI config and initialize cadence PDI */
1426 intel_pdi_init(sdw
, &config
);
1427 ret
= sdw_cdns_pdi_init(cdns
, config
);
1431 intel_pdi_ch_update(sdw
);
1433 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1435 dev_err(dev
, "cannot enable interrupts\n");
1440 * follow recommended programming flows to avoid timeouts when
1444 intel_shim_sync_arm(sdw
);
1446 ret
= sdw_cdns_init(cdns
);
1448 dev_err(dev
, "unable to initialize Cadence IP\n");
1452 ret
= sdw_cdns_exit_reset(cdns
);
1454 dev_err(dev
, "unable to exit bus reset sequence\n");
1459 ret
= intel_shim_sync_go(sdw
);
1461 dev_err(dev
, "sync go failed: %d\n", ret
);
1467 ret
= intel_register_dai(sdw
);
1469 dev_err(dev
, "DAI registration failed: %d\n", ret
);
1470 snd_soc_unregister_component(dev
);
1474 intel_debugfs_init(sdw
);
1476 /* Enable runtime PM */
1477 if (!(link_flags
& SDW_INTEL_MASTER_DISABLE_PM_RUNTIME
)) {
1478 pm_runtime_set_autosuspend_delay(dev
,
1479 INTEL_MASTER_SUSPEND_DELAY_MS
);
1480 pm_runtime_use_autosuspend(dev
);
1481 pm_runtime_mark_last_busy(dev
);
1483 pm_runtime_set_active(dev
);
1484 pm_runtime_enable(dev
);
1487 clock_stop_quirks
= sdw
->link_res
->clock_stop_quirks
;
1488 if (clock_stop_quirks
& SDW_INTEL_CLK_STOP_NOT_ALLOWED
) {
1490 * To keep the clock running we need to prevent
1491 * pm_runtime suspend from happening by increasing the
1493 * This quirk is specified by the parent PCI device in
1494 * case of specific latency requirements. It will have
1495 * no effect if pm_runtime is disabled by the user via
1496 * a module parameter for testing purposes.
1498 pm_runtime_get_noresume(dev
);
1502 * The runtime PM status of Slave devices is "Unsupported"
1503 * until they report as ATTACHED. If they don't, e.g. because
1504 * there are no Slave devices populated or if the power-on is
1505 * delayed or dependent on a power switch, the Master will
1506 * remain active and prevent its parent from suspending.
1508 * Conditionally force the pm_runtime core to re-evaluate the
1509 * Master status in the absence of any Slave activity. A quirk
1510 * is provided to e.g. deal with Slaves that may be powered on
1511 * with a delay. A more complete solution would require the
1512 * definition of Master properties.
1514 if (!(link_flags
& SDW_INTEL_MASTER_DISABLE_PM_RUNTIME_IDLE
))
1515 pm_runtime_idle(dev
);
1520 sdw_cdns_enable_interrupt(cdns
, false);
1525 static int intel_master_remove(struct platform_device
*pdev
)
1527 struct device
*dev
= &pdev
->dev
;
1528 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1529 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1530 struct sdw_bus
*bus
= &cdns
->bus
;
1533 * Since pm_runtime is already disabled, we don't decrease
1534 * the refcount when the clock_stop_quirk is
1535 * SDW_INTEL_CLK_STOP_NOT_ALLOWED
1537 if (!bus
->prop
.hw_disabled
) {
1538 intel_debugfs_exit(sdw
);
1539 sdw_cdns_enable_interrupt(cdns
, false);
1540 snd_soc_unregister_component(dev
);
1542 sdw_bus_master_delete(bus
);
1547 int intel_master_process_wakeen_event(struct platform_device
*pdev
)
1549 struct device
*dev
= &pdev
->dev
;
1550 struct sdw_intel
*sdw
;
1551 struct sdw_bus
*bus
;
1555 sdw
= platform_get_drvdata(pdev
);
1556 bus
= &sdw
->cdns
.bus
;
1558 if (bus
->prop
.hw_disabled
) {
1559 dev_dbg(dev
, "SoundWire master %d is disabled, ignoring\n", bus
->link_id
);
1563 shim
= sdw
->link_res
->shim
;
1564 wake_sts
= intel_readw(shim
, SDW_SHIM_WAKESTS
);
1566 if (!(wake_sts
& BIT(sdw
->instance
)))
1569 /* disable WAKEEN interrupt ASAP to prevent interrupt flood */
1570 intel_shim_wake(sdw
, false);
1573 * resume the Master, which will generate a bus reset and result in
1574 * Slaves re-attaching and be re-enumerated. The SoundWire physical
1575 * device which generated the wake will trigger an interrupt, which
1576 * will in turn cause the corresponding Linux Slave device to be
1577 * resumed and the Slave codec driver to check the status.
1579 pm_request_resume(dev
);
1588 static int __maybe_unused
intel_suspend(struct device
*dev
)
1590 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1591 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1592 struct sdw_bus
*bus
= &cdns
->bus
;
1593 u32 clock_stop_quirks
;
1596 if (bus
->prop
.hw_disabled
) {
1597 dev_dbg(dev
, "SoundWire master %d is disabled, ignoring\n",
1602 if (pm_runtime_suspended(dev
)) {
1603 dev_dbg(dev
, "%s: pm_runtime status: suspended\n", __func__
);
1605 clock_stop_quirks
= sdw
->link_res
->clock_stop_quirks
;
1607 if ((clock_stop_quirks
& SDW_INTEL_CLK_STOP_BUS_RESET
||
1608 !clock_stop_quirks
) &&
1609 !pm_runtime_suspended(dev
->parent
)) {
1612 * if we've enabled clock stop, and the parent
1613 * is still active, disable shim wake. The
1614 * SHIM registers are not accessible if the
1615 * parent is already pm_runtime suspended so
1616 * it's too late to change that configuration
1619 intel_shim_wake(sdw
, false);
1625 ret
= sdw_cdns_enable_interrupt(cdns
, false);
1627 dev_err(dev
, "cannot disable interrupts on suspend\n");
1631 ret
= intel_link_power_down(sdw
);
1633 dev_err(dev
, "Link power down failed: %d", ret
);
1637 intel_shim_wake(sdw
, false);
1642 static int __maybe_unused
intel_suspend_runtime(struct device
*dev
)
1644 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1645 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1646 struct sdw_bus
*bus
= &cdns
->bus
;
1647 u32 clock_stop_quirks
;
1650 if (bus
->prop
.hw_disabled
) {
1651 dev_dbg(dev
, "SoundWire master %d is disabled, ignoring\n",
1656 clock_stop_quirks
= sdw
->link_res
->clock_stop_quirks
;
1658 if (clock_stop_quirks
& SDW_INTEL_CLK_STOP_TEARDOWN
) {
1660 ret
= sdw_cdns_enable_interrupt(cdns
, false);
1662 dev_err(dev
, "cannot disable interrupts on suspend\n");
1666 ret
= intel_link_power_down(sdw
);
1668 dev_err(dev
, "Link power down failed: %d", ret
);
1672 intel_shim_wake(sdw
, false);
1674 } else if (clock_stop_quirks
& SDW_INTEL_CLK_STOP_BUS_RESET
||
1675 !clock_stop_quirks
) {
1676 ret
= sdw_cdns_clock_stop(cdns
, true);
1678 dev_err(dev
, "cannot enable clock stop on suspend\n");
1682 ret
= sdw_cdns_enable_interrupt(cdns
, false);
1684 dev_err(dev
, "cannot disable interrupts on suspend\n");
1688 ret
= intel_link_power_down(sdw
);
1690 dev_err(dev
, "Link power down failed: %d", ret
);
1694 intel_shim_wake(sdw
, true);
1696 dev_err(dev
, "%s clock_stop_quirks %x unsupported\n",
1697 __func__
, clock_stop_quirks
);
1704 static int __maybe_unused
intel_resume(struct device
*dev
)
1706 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1707 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1708 struct sdw_bus
*bus
= &cdns
->bus
;
1713 if (bus
->prop
.hw_disabled
) {
1714 dev_dbg(dev
, "SoundWire master %d is disabled, ignoring\n",
1719 link_flags
= md_flags
>> (bus
->link_id
* 8);
1720 multi_link
= !(link_flags
& SDW_INTEL_MASTER_DISABLE_MULTI_LINK
);
1722 if (pm_runtime_suspended(dev
)) {
1723 dev_dbg(dev
, "%s: pm_runtime status was suspended, forcing active\n", __func__
);
1725 /* follow required sequence from runtime_pm.rst */
1726 pm_runtime_disable(dev
);
1727 pm_runtime_set_active(dev
);
1728 pm_runtime_mark_last_busy(dev
);
1729 pm_runtime_enable(dev
);
1731 link_flags
= md_flags
>> (bus
->link_id
* 8);
1733 if (!(link_flags
& SDW_INTEL_MASTER_DISABLE_PM_RUNTIME_IDLE
))
1734 pm_runtime_idle(dev
);
1737 ret
= intel_init(sdw
);
1739 dev_err(dev
, "%s failed: %d", __func__
, ret
);
1744 * make sure all Slaves are tagged as UNATTACHED and provide
1745 * reason for reinitialization
1747 sdw_clear_slave_status(bus
, SDW_UNATTACH_REQUEST_MASTER_RESET
);
1749 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1751 dev_err(dev
, "cannot enable interrupts during resume\n");
1756 * follow recommended programming flows to avoid timeouts when
1760 intel_shim_sync_arm(sdw
);
1762 ret
= sdw_cdns_init(&sdw
->cdns
);
1764 dev_err(dev
, "unable to initialize Cadence IP during resume\n");
1768 ret
= sdw_cdns_exit_reset(cdns
);
1770 dev_err(dev
, "unable to exit bus reset sequence during resume\n");
1775 ret
= intel_shim_sync_go(sdw
);
1777 dev_err(dev
, "sync go failed during resume\n");
1783 * after system resume, the pm_runtime suspend() may kick in
1784 * during the enumeration, before any children device force the
1785 * master device to remain active. Using pm_runtime_get()
1786 * routines is not really possible, since it'd prevent the
1787 * master from suspending.
1788 * A reasonable compromise is to update the pm_runtime
1789 * counters and delay the pm_runtime suspend by several
1790 * seconds, by when all enumeration should be complete.
1792 pm_runtime_mark_last_busy(dev
);
1797 static int __maybe_unused
intel_resume_runtime(struct device
*dev
)
1799 struct sdw_cdns
*cdns
= dev_get_drvdata(dev
);
1800 struct sdw_intel
*sdw
= cdns_to_intel(cdns
);
1801 struct sdw_bus
*bus
= &cdns
->bus
;
1802 u32 clock_stop_quirks
;
1809 if (bus
->prop
.hw_disabled
) {
1810 dev_dbg(dev
, "SoundWire master %d is disabled, ignoring\n",
1815 link_flags
= md_flags
>> (bus
->link_id
* 8);
1816 multi_link
= !(link_flags
& SDW_INTEL_MASTER_DISABLE_MULTI_LINK
);
1818 clock_stop_quirks
= sdw
->link_res
->clock_stop_quirks
;
1820 if (clock_stop_quirks
& SDW_INTEL_CLK_STOP_TEARDOWN
) {
1821 ret
= intel_init(sdw
);
1823 dev_err(dev
, "%s failed: %d", __func__
, ret
);
1828 * make sure all Slaves are tagged as UNATTACHED and provide
1829 * reason for reinitialization
1831 sdw_clear_slave_status(bus
, SDW_UNATTACH_REQUEST_MASTER_RESET
);
1833 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1835 dev_err(dev
, "cannot enable interrupts during resume\n");
1840 * follow recommended programming flows to avoid
1841 * timeouts when gsync is enabled
1844 intel_shim_sync_arm(sdw
);
1846 ret
= sdw_cdns_init(&sdw
->cdns
);
1848 dev_err(dev
, "unable to initialize Cadence IP during resume\n");
1852 ret
= sdw_cdns_exit_reset(cdns
);
1854 dev_err(dev
, "unable to exit bus reset sequence during resume\n");
1859 ret
= intel_shim_sync_go(sdw
);
1861 dev_err(dev
, "sync go failed during resume\n");
1865 } else if (clock_stop_quirks
& SDW_INTEL_CLK_STOP_BUS_RESET
) {
1866 ret
= intel_init(sdw
);
1868 dev_err(dev
, "%s failed: %d", __func__
, ret
);
1873 * An exception condition occurs for the CLK_STOP_BUS_RESET
1874 * case if one or more masters remain active. In this condition,
1875 * all the masters are powered on for they are in the same power
1876 * domain. Master can preserve its context for clock stop0, so
1877 * there is no need to clear slave status and reset bus.
1879 clock_stop0
= sdw_cdns_is_clock_stop(&sdw
->cdns
);
1884 * make sure all Slaves are tagged as UNATTACHED and
1885 * provide reason for reinitialization
1888 status
= SDW_UNATTACH_REQUEST_MASTER_RESET
;
1889 sdw_clear_slave_status(bus
, status
);
1891 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1893 dev_err(dev
, "cannot enable interrupts during resume\n");
1898 * follow recommended programming flows to avoid
1899 * timeouts when gsync is enabled
1902 intel_shim_sync_arm(sdw
);
1905 * Re-initialize the IP since it was powered-off
1907 sdw_cdns_init(&sdw
->cdns
);
1910 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1912 dev_err(dev
, "cannot enable interrupts during resume\n");
1917 ret
= sdw_cdns_clock_restart(cdns
, !clock_stop0
);
1919 dev_err(dev
, "unable to restart clock during resume\n");
1924 ret
= sdw_cdns_exit_reset(cdns
);
1926 dev_err(dev
, "unable to exit bus reset sequence during resume\n");
1931 ret
= intel_shim_sync_go(sdw
);
1933 dev_err(sdw
->cdns
.dev
, "sync go failed during resume\n");
1938 } else if (!clock_stop_quirks
) {
1940 clock_stop0
= sdw_cdns_is_clock_stop(&sdw
->cdns
);
1942 dev_err(dev
, "%s invalid configuration, clock was not stopped", __func__
);
1944 ret
= intel_init(sdw
);
1946 dev_err(dev
, "%s failed: %d", __func__
, ret
);
1950 ret
= sdw_cdns_enable_interrupt(cdns
, true);
1952 dev_err(dev
, "cannot enable interrupts during resume\n");
1956 ret
= sdw_cdns_clock_restart(cdns
, false);
1958 dev_err(dev
, "unable to resume master during resume\n");
1962 dev_err(dev
, "%s clock_stop_quirks %x unsupported\n",
1963 __func__
, clock_stop_quirks
);
1970 static const struct dev_pm_ops intel_pm
= {
1971 SET_SYSTEM_SLEEP_PM_OPS(intel_suspend
, intel_resume
)
1972 SET_RUNTIME_PM_OPS(intel_suspend_runtime
, intel_resume_runtime
, NULL
)
1975 static struct platform_driver sdw_intel_drv
= {
1976 .probe
= intel_master_probe
,
1977 .remove
= intel_master_remove
,
1979 .name
= "intel-sdw",
1984 module_platform_driver(sdw_intel_drv
);
1986 MODULE_LICENSE("Dual BSD/GPL");
1987 MODULE_ALIAS("platform:intel-sdw");
1988 MODULE_DESCRIPTION("Intel Soundwire Master Driver");