mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-14 22:44:27 +08:00
f2183847cf
struct i2c_driver::probe_new() is about to go away. Since I converted all drivers below drivers/media use struct i2c_driver::probe, three more drivers were added in the following commits that use .probe_new(): commit6363db1c9d
("media: i2c: add DS90UB953 driver") commitc158d0d4ff
("media: i2c: add DS90UB913 driver") commitafe267f2d3
("media: i2c: add DS90UB960 driver") Switch these driver to use the probe callback. Signed-off-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de> Signed-off-by: Hans Verkuil <hverkuil-cisco@xs4all.nl> [hverkuil: use proper commit description style]
4052 lines
100 KiB
C
4052 lines
100 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Driver for the Texas Instruments DS90UB960-Q1 video deserializer
|
|
*
|
|
* Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
|
|
* Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
|
|
*/
|
|
|
|
/*
|
|
* (Possible) TODOs:
|
|
*
|
|
* - PM for serializer and remote peripherals. We need to manage:
|
|
* - VPOC
|
|
* - Power domain? Regulator? Somehow any remote device should be able to
|
|
* cause the VPOC to be turned on.
|
|
* - Link between the deserializer and the serializer
|
|
* - Related to VPOC management. We probably always want to turn on the VPOC
|
|
* and then enable the link.
|
|
* - Serializer's services: i2c, gpios, power
|
|
* - The serializer needs to resume before the remote peripherals can
|
|
* e.g. use the i2c.
|
|
* - How to handle gpios? Reserving a gpio essentially keeps the provider
|
|
* (serializer) always powered on.
|
|
* - Do we need a new bus for the FPD-Link? At the moment the serializers
|
|
* are children of the same i2c-adapter where the deserializer resides.
|
|
* - i2c-atr could be made embeddable instead of allocatable.
|
|
*/
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/fwnode.h>
|
|
#include <linux/gpio/consumer.h>
|
|
#include <linux/i2c-atr.h>
|
|
#include <linux/i2c.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/property.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/regulator/consumer.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/workqueue.h>
|
|
|
|
#include <media/i2c/ds90ub9xx.h>
|
|
#include <media/mipi-csi2.h>
|
|
#include <media/v4l2-ctrls.h>
|
|
#include <media/v4l2-event.h>
|
|
#include <media/v4l2-fwnode.h>
|
|
#include <media/v4l2-subdev.h>
|
|
|
|
#define MHZ(v) ((u32)((v) * 1000000U))
|
|
|
|
#define UB960_POLL_TIME_MS 500
|
|
|
|
#define UB960_MAX_RX_NPORTS 4
|
|
#define UB960_MAX_TX_NPORTS 2
|
|
#define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
|
|
|
|
#define UB960_MAX_PORT_ALIASES 8
|
|
|
|
#define UB960_NUM_BC_GPIOS 4
|
|
|
|
/*
|
|
* Register map
|
|
*
|
|
* 0x00-0x32 Shared (UB960_SR)
|
|
* 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
|
|
* 0x4c Shared (UB960_SR)
|
|
* 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
|
|
* 0xb0-0xbf Shared (UB960_SR)
|
|
* 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
|
|
* 0xf0-0xf5 Shared (UB960_SR)
|
|
* 0xf8-0xfb Shared (UB960_SR)
|
|
* All others Reserved
|
|
*
|
|
* Register prefixes:
|
|
* UB960_SR_* = Shared register
|
|
* UB960_RR_* = FPD-Link RX, per-port paged register
|
|
* UB960_TR_* = CSI-2 TX, per-port paged register
|
|
* UB960_XR_* = Reserved register
|
|
* UB960_IR_* = Indirect register
|
|
*/
|
|
|
|
#define UB960_SR_I2C_DEV_ID 0x00
|
|
#define UB960_SR_RESET 0x01
|
|
#define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
|
|
#define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
|
|
#define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
|
|
|
|
#define UB960_SR_GEN_CONFIG 0x02
|
|
#define UB960_SR_REV_MASK 0x03
|
|
#define UB960_SR_DEVICE_STS 0x04
|
|
#define UB960_SR_PAR_ERR_THOLD_HI 0x05
|
|
#define UB960_SR_PAR_ERR_THOLD_LO 0x06
|
|
#define UB960_SR_BCC_WDOG_CTL 0x07
|
|
#define UB960_SR_I2C_CTL1 0x08
|
|
#define UB960_SR_I2C_CTL2 0x09
|
|
#define UB960_SR_SCL_HIGH_TIME 0x0a
|
|
#define UB960_SR_SCL_LOW_TIME 0x0b
|
|
#define UB960_SR_RX_PORT_CTL 0x0c
|
|
#define UB960_SR_IO_CTL 0x0d
|
|
#define UB960_SR_GPIO_PIN_STS 0x0e
|
|
#define UB960_SR_GPIO_INPUT_CTL 0x0f
|
|
#define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
|
|
|
|
#define UB960_SR_FS_CTL 0x18
|
|
#define UB960_SR_FS_HIGH_TIME_1 0x19
|
|
#define UB960_SR_FS_HIGH_TIME_0 0x1a
|
|
#define UB960_SR_FS_LOW_TIME_1 0x1b
|
|
#define UB960_SR_FS_LOW_TIME_0 0x1c
|
|
#define UB960_SR_MAX_FRM_HI 0x1d
|
|
#define UB960_SR_MAX_FRM_LO 0x1e
|
|
#define UB960_SR_CSI_PLL_CTL 0x1f
|
|
|
|
#define UB960_SR_FWD_CTL1 0x20
|
|
#define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
|
|
|
|
#define UB960_SR_FWD_CTL2 0x21
|
|
#define UB960_SR_FWD_STS 0x22
|
|
|
|
#define UB960_SR_INTERRUPT_CTL 0x23
|
|
#define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
|
|
#define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
|
|
#define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
|
|
|
|
#define UB960_SR_INTERRUPT_STS 0x24
|
|
#define UB960_SR_INTERRUPT_STS_INT BIT(7)
|
|
#define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
|
|
#define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
|
|
|
|
#define UB960_SR_TS_CONFIG 0x25
|
|
#define UB960_SR_TS_CONTROL 0x26
|
|
#define UB960_SR_TS_LINE_HI 0x27
|
|
#define UB960_SR_TS_LINE_LO 0x28
|
|
#define UB960_SR_TS_STATUS 0x29
|
|
#define UB960_SR_TIMESTAMP_P0_HI 0x2a
|
|
#define UB960_SR_TIMESTAMP_P0_LO 0x2b
|
|
#define UB960_SR_TIMESTAMP_P1_HI 0x2c
|
|
#define UB960_SR_TIMESTAMP_P1_LO 0x2d
|
|
|
|
#define UB960_SR_CSI_PORT_SEL 0x32
|
|
|
|
#define UB960_TR_CSI_CTL 0x33
|
|
#define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
|
|
#define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
|
|
|
|
#define UB960_TR_CSI_CTL2 0x34
|
|
#define UB960_TR_CSI_STS 0x35
|
|
#define UB960_TR_CSI_TX_ICR 0x36
|
|
|
|
#define UB960_TR_CSI_TX_ISR 0x37
|
|
#define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
|
|
#define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
|
|
|
|
#define UB960_TR_CSI_TEST_CTL 0x38
|
|
#define UB960_TR_CSI_TEST_PATT_HI 0x39
|
|
#define UB960_TR_CSI_TEST_PATT_LO 0x3a
|
|
|
|
#define UB960_XR_SFILTER_CFG 0x41
|
|
#define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
|
|
#define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
|
|
|
|
#define UB960_XR_AEQ_CTL1 0x42
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
|
|
(UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
|
|
|
|
#define UB960_XR_AEQ_ERR_THOLD 0x43
|
|
|
|
#define UB960_RR_BCC_ERR_CTL 0x46
|
|
#define UB960_RR_BCC_STATUS 0x47
|
|
#define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
|
|
#define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
|
|
#define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
|
|
#define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
|
|
#define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
|
|
#define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
|
|
#define UB960_RR_BCC_STATUS_ERROR_MASK \
|
|
(UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
|
|
UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
|
|
UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
|
|
|
|
#define UB960_RR_FPD3_CAP 0x4a
|
|
#define UB960_RR_RAW_EMBED_DTYPE 0x4b
|
|
#define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
|
|
|
|
#define UB960_SR_FPD3_PORT_SEL 0x4c
|
|
|
|
#define UB960_RR_RX_PORT_STS1 0x4d
|
|
#define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
|
|
#define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
|
|
#define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
|
|
#define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
|
|
#define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
|
|
#define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
|
|
#define UB960_RR_RX_PORT_STS1_ERROR_MASK \
|
|
(UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
|
|
UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
|
|
UB960_RR_RX_PORT_STS1_PARITY_ERROR)
|
|
|
|
#define UB960_RR_RX_PORT_STS2 0x4e
|
|
#define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
|
|
#define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
|
|
#define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
|
|
#define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
|
|
#define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
|
|
#define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
|
|
#define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
|
|
#define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
|
|
#define UB960_RR_RX_PORT_STS2_ERROR_MASK \
|
|
UB960_RR_RX_PORT_STS2_BUFFER_ERROR
|
|
|
|
#define UB960_RR_RX_FREQ_HIGH 0x4f
|
|
#define UB960_RR_RX_FREQ_LOW 0x50
|
|
#define UB960_RR_SENSOR_STS_0 0x51
|
|
#define UB960_RR_SENSOR_STS_1 0x52
|
|
#define UB960_RR_SENSOR_STS_2 0x53
|
|
#define UB960_RR_SENSOR_STS_3 0x54
|
|
#define UB960_RR_RX_PAR_ERR_HI 0x55
|
|
#define UB960_RR_RX_PAR_ERR_LO 0x56
|
|
#define UB960_RR_BIST_ERR_COUNT 0x57
|
|
|
|
#define UB960_RR_BCC_CONFIG 0x58
|
|
#define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
|
|
#define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
|
|
|
|
#define UB960_RR_DATAPATH_CTL1 0x59
|
|
#define UB960_RR_DATAPATH_CTL2 0x5a
|
|
#define UB960_RR_SER_ID 0x5b
|
|
#define UB960_RR_SER_ALIAS_ID 0x5c
|
|
|
|
/* For these two register sets: n < UB960_MAX_PORT_ALIASES */
|
|
#define UB960_RR_SLAVE_ID(n) (0x5d + (n))
|
|
#define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
|
|
|
|
#define UB960_RR_PORT_CONFIG 0x6d
|
|
#define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
|
|
|
|
#define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
|
|
#define UB960_RR_RAW10_ID 0x70
|
|
#define UB960_RR_RAW10_ID_VC_SHIFT 6
|
|
#define UB960_RR_RAW10_ID_DT_SHIFT 0
|
|
|
|
#define UB960_RR_RAW12_ID 0x71
|
|
#define UB960_RR_CSI_VC_MAP 0x72
|
|
#define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
|
|
|
|
#define UB960_RR_LINE_COUNT_HI 0x73
|
|
#define UB960_RR_LINE_COUNT_LO 0x74
|
|
#define UB960_RR_LINE_LEN_1 0x75
|
|
#define UB960_RR_LINE_LEN_0 0x76
|
|
#define UB960_RR_FREQ_DET_CTL 0x77
|
|
#define UB960_RR_MAILBOX_1 0x78
|
|
#define UB960_RR_MAILBOX_2 0x79
|
|
|
|
#define UB960_RR_CSI_RX_STS 0x7a
|
|
#define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
|
|
#define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
|
|
#define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
|
|
#define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
|
|
#define UB960_RR_CSI_RX_STS_ERROR_MASK \
|
|
(UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
|
|
UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
|
|
|
|
#define UB960_RR_CSI_ERR_COUNTER 0x7b
|
|
#define UB960_RR_PORT_CONFIG2 0x7c
|
|
#define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
|
|
#define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
|
|
|
|
#define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
|
|
#define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
|
|
|
|
#define UB960_RR_PORT_PASS_CTL 0x7d
|
|
#define UB960_RR_SEN_INT_RISE_CTL 0x7e
|
|
#define UB960_RR_SEN_INT_FALL_CTL 0x7f
|
|
|
|
#define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
|
|
|
|
#define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
|
|
|
|
#define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
|
|
|
|
#define UB960_SR_IND_ACC_CTL 0xb0
|
|
#define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
|
|
|
|
#define UB960_SR_IND_ACC_ADDR 0xb1
|
|
#define UB960_SR_IND_ACC_DATA 0xb2
|
|
#define UB960_SR_BIST_CONTROL 0xb3
|
|
#define UB960_SR_MODE_IDX_STS 0xb8
|
|
#define UB960_SR_LINK_ERROR_COUNT 0xb9
|
|
#define UB960_SR_FPD3_ENC_CTL 0xba
|
|
#define UB960_SR_FV_MIN_TIME 0xbc
|
|
#define UB960_SR_GPIO_PD_CTL 0xbe
|
|
|
|
#define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
|
|
#define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
|
|
|
|
#define UB960_RR_PORT_DEBUG 0xd0
|
|
#define UB960_RR_AEQ_CTL2 0xd2
|
|
#define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
|
|
|
|
#define UB960_RR_AEQ_STATUS 0xd3
|
|
#define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
|
|
#define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
|
|
|
|
#define UB960_RR_AEQ_BYPASS 0xd4
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
|
|
#define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
|
|
|
|
#define UB960_RR_AEQ_MIN_MAX 0xd5
|
|
#define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
|
|
#define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
|
|
|
|
#define UB960_RR_SFILTER_STS_0 0xd6
|
|
#define UB960_RR_SFILTER_STS_1 0xd7
|
|
#define UB960_RR_PORT_ICR_HI 0xd8
|
|
#define UB960_RR_PORT_ICR_LO 0xd9
|
|
#define UB960_RR_PORT_ISR_HI 0xda
|
|
#define UB960_RR_PORT_ISR_LO 0xdb
|
|
#define UB960_RR_FC_GPIO_STS 0xdc
|
|
#define UB960_RR_FC_GPIO_ICR 0xdd
|
|
#define UB960_RR_SEN_INT_RISE_STS 0xde
|
|
#define UB960_RR_SEN_INT_FALL_STS 0xdf
|
|
|
|
#define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
|
|
|
|
#define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
|
|
#define UB960_SR_FPD3_RX_ID_LEN 6
|
|
|
|
#define UB960_SR_I2C_RX_ID(n) (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
|
|
|
|
/* Indirect register blocks */
|
|
#define UB960_IND_TARGET_PAT_GEN 0x00
|
|
#define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
|
|
#define UB960_IND_TARGET_CSI_CSIPLL_REG_1 0x92 /* UB9702 */
|
|
#define UB960_IND_TARGET_CSI_ANA 0x07
|
|
|
|
/* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
|
|
|
|
#define UB960_IR_PGEN_CTL 0x01
|
|
#define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
|
|
|
|
#define UB960_IR_PGEN_CFG 0x02
|
|
#define UB960_IR_PGEN_CSI_DI 0x03
|
|
#define UB960_IR_PGEN_LINE_SIZE1 0x04
|
|
#define UB960_IR_PGEN_LINE_SIZE0 0x05
|
|
#define UB960_IR_PGEN_BAR_SIZE1 0x06
|
|
#define UB960_IR_PGEN_BAR_SIZE0 0x07
|
|
#define UB960_IR_PGEN_ACT_LPF1 0x08
|
|
#define UB960_IR_PGEN_ACT_LPF0 0x09
|
|
#define UB960_IR_PGEN_TOT_LPF1 0x0a
|
|
#define UB960_IR_PGEN_TOT_LPF0 0x0b
|
|
#define UB960_IR_PGEN_LINE_PD1 0x0c
|
|
#define UB960_IR_PGEN_LINE_PD0 0x0d
|
|
#define UB960_IR_PGEN_VBP 0x0e
|
|
#define UB960_IR_PGEN_VFP 0x0f
|
|
#define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
|
|
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
|
|
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
|
|
|
|
/* EQ related */
|
|
|
|
#define UB960_MIN_AEQ_STROBE_POS -7
|
|
#define UB960_MAX_AEQ_STROBE_POS 7
|
|
|
|
#define UB960_MANUAL_STROBE_EXTRA_DELAY 6
|
|
|
|
#define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
|
|
#define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
|
|
#define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
|
|
|
|
#define UB960_MIN_EQ_LEVEL 0
|
|
#define UB960_MAX_EQ_LEVEL 14
|
|
#define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
|
|
|
|
struct ub960_hw_data {
|
|
const char *model;
|
|
u8 num_rxports;
|
|
u8 num_txports;
|
|
bool is_ub9702;
|
|
bool is_fpdlink4;
|
|
};
|
|
|
|
enum ub960_rxport_mode {
|
|
RXPORT_MODE_RAW10 = 0,
|
|
RXPORT_MODE_RAW12_HF = 1,
|
|
RXPORT_MODE_RAW12_LF = 2,
|
|
RXPORT_MODE_CSI2_SYNC = 3,
|
|
RXPORT_MODE_CSI2_ASYNC = 4,
|
|
RXPORT_MODE_LAST = RXPORT_MODE_CSI2_ASYNC,
|
|
};
|
|
|
|
enum ub960_rxport_cdr {
|
|
RXPORT_CDR_FPD3 = 0,
|
|
RXPORT_CDR_FPD4 = 1,
|
|
RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
|
|
};
|
|
|
|
struct ub960_rxport {
|
|
struct ub960_data *priv;
|
|
u8 nport; /* RX port number, and index in priv->rxport[] */
|
|
|
|
struct {
|
|
struct v4l2_subdev *sd;
|
|
u16 pad;
|
|
struct fwnode_handle *ep_fwnode;
|
|
} source;
|
|
|
|
/* Serializer */
|
|
struct {
|
|
struct fwnode_handle *fwnode;
|
|
struct i2c_client *client;
|
|
unsigned short alias; /* I2C alias (lower 7 bits) */
|
|
struct ds90ub9xx_platform_data pdata;
|
|
} ser;
|
|
|
|
enum ub960_rxport_mode rx_mode;
|
|
enum ub960_rxport_cdr cdr_mode;
|
|
|
|
u8 lv_fv_pol; /* LV and FV polarities */
|
|
|
|
struct regulator *vpoc;
|
|
|
|
/* EQ settings */
|
|
struct {
|
|
bool manual_eq;
|
|
|
|
s8 strobe_pos;
|
|
|
|
union {
|
|
struct {
|
|
u8 eq_level_min;
|
|
u8 eq_level_max;
|
|
} aeq;
|
|
|
|
struct {
|
|
u8 eq_level;
|
|
} manual;
|
|
};
|
|
} eq;
|
|
|
|
const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
|
|
};
|
|
|
|
struct ub960_asd {
|
|
struct v4l2_async_subdev base;
|
|
struct ub960_rxport *rxport;
|
|
};
|
|
|
|
static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_subdev *asd)
|
|
{
|
|
return container_of(asd, struct ub960_asd, base);
|
|
}
|
|
|
|
struct ub960_txport {
|
|
struct ub960_data *priv;
|
|
u8 nport; /* TX port number, and index in priv->txport[] */
|
|
|
|
u32 num_data_lanes;
|
|
};
|
|
|
|
struct ub960_data {
|
|
const struct ub960_hw_data *hw_data;
|
|
struct i2c_client *client; /* for shared local registers */
|
|
struct regmap *regmap;
|
|
|
|
/* lock for register access */
|
|
struct mutex reg_lock;
|
|
|
|
struct clk *refclk;
|
|
|
|
struct regulator *vddio;
|
|
|
|
struct gpio_desc *pd_gpio;
|
|
struct delayed_work poll_work;
|
|
struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
|
|
struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
|
|
|
|
struct v4l2_subdev sd;
|
|
struct media_pad pads[UB960_MAX_NPORTS];
|
|
|
|
struct v4l2_ctrl_handler ctrl_handler;
|
|
struct v4l2_async_notifier notifier;
|
|
|
|
u32 tx_data_rate; /* Nominal data rate (Gb/s) */
|
|
s64 tx_link_freq[1];
|
|
|
|
struct i2c_atr *atr;
|
|
|
|
struct {
|
|
u8 rxport;
|
|
u8 txport;
|
|
u8 indirect_target;
|
|
} reg_current;
|
|
|
|
bool streaming;
|
|
|
|
u8 stored_fwd_ctl;
|
|
|
|
u64 stream_enable_mask[UB960_MAX_NPORTS];
|
|
|
|
/* These are common to all ports */
|
|
struct {
|
|
bool manual;
|
|
|
|
s8 min;
|
|
s8 max;
|
|
} strobe;
|
|
};
|
|
|
|
static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
|
|
{
|
|
return container_of(sd, struct ub960_data, sd);
|
|
}
|
|
|
|
static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
|
|
{
|
|
return pad < priv->hw_data->num_rxports;
|
|
}
|
|
|
|
static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
|
|
{
|
|
return pad >= priv->hw_data->num_rxports;
|
|
}
|
|
|
|
static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
|
|
{
|
|
if (ub960_pad_is_sink(priv, pad))
|
|
return pad;
|
|
else
|
|
return pad - priv->hw_data->num_rxports;
|
|
}
|
|
|
|
struct ub960_format_info {
|
|
u32 code;
|
|
u32 bpp;
|
|
u8 datatype;
|
|
bool meta;
|
|
};
|
|
|
|
static const struct ub960_format_info ub960_formats[] = {
|
|
{ .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
|
|
{ .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
};
|
|
|
|
static const struct ub960_format_info *ub960_find_format(u32 code)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
|
|
if (ub960_formats[i].code == code)
|
|
return &ub960_formats[i];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* Basic device access
|
|
*/
|
|
|
|
static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
__be16 __v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = be16_to_cpu(__v);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.rxport == nport)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
|
|
(nport << 4) | BIT(nport));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
|
|
nport, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.rxport = nport;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u16 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
__be16 __v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = be16_to_cpu(__v);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_select(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.txport == nport)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
|
|
(nport << 4) | BIT(nport));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
|
|
nport, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.txport = nport;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.indirect_target == block)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
|
|
__func__, block, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.indirect_target = block;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
|
|
val);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* I2C-ATR (address translator)
|
|
*/
|
|
|
|
static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
|
|
const struct i2c_client *client, u16 alias)
|
|
{
|
|
struct ub960_data *priv = i2c_atr_get_driver_data(atr);
|
|
struct ub960_rxport *rxport = priv->rxports[chan_id];
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int reg_idx;
|
|
|
|
for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
|
|
if (!rxport->aliased_clients[reg_idx])
|
|
break;
|
|
}
|
|
|
|
if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
|
|
dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
|
|
return -EADDRNOTAVAIL;
|
|
}
|
|
|
|
rxport->aliased_clients[reg_idx] = client;
|
|
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
|
|
client->addr << 1);
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
|
|
alias << 1);
|
|
|
|
dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
|
|
rxport->nport, client->addr, alias, reg_idx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
|
|
const struct i2c_client *client)
|
|
{
|
|
struct ub960_data *priv = i2c_atr_get_driver_data(atr);
|
|
struct ub960_rxport *rxport = priv->rxports[chan_id];
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int reg_idx;
|
|
|
|
for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
|
|
if (rxport->aliased_clients[reg_idx] == client)
|
|
break;
|
|
}
|
|
|
|
if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
|
|
dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
|
|
rxport->nport, client->addr);
|
|
return;
|
|
}
|
|
|
|
rxport->aliased_clients[reg_idx] = NULL;
|
|
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
|
|
|
|
dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
|
|
client->addr, reg_idx);
|
|
}
|
|
|
|
static const struct i2c_atr_ops ub960_atr_ops = {
|
|
.attach_client = ub960_atr_attach_client,
|
|
.detach_client = ub960_atr_detach_client,
|
|
};
|
|
|
|
static int ub960_init_atr(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct i2c_adapter *parent_adap = priv->client->adapter;
|
|
|
|
priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
|
|
priv->hw_data->num_rxports);
|
|
if (IS_ERR(priv->atr))
|
|
return PTR_ERR(priv->atr);
|
|
|
|
i2c_atr_set_driver_data(priv->atr, priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_uninit_atr(struct ub960_data *priv)
|
|
{
|
|
i2c_atr_delete(priv->atr);
|
|
priv->atr = NULL;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* TX ports
|
|
*/
|
|
|
|
static int ub960_parse_dt_txport(struct ub960_data *priv,
|
|
struct fwnode_handle *ep_fwnode,
|
|
u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_fwnode_endpoint vep = {};
|
|
struct ub960_txport *txport;
|
|
int ret;
|
|
|
|
txport = kzalloc(sizeof(*txport), GFP_KERNEL);
|
|
if (!txport)
|
|
return -ENOMEM;
|
|
|
|
txport->priv = priv;
|
|
txport->nport = nport;
|
|
|
|
vep.bus_type = V4L2_MBUS_CSI2_DPHY;
|
|
ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
|
|
if (ret) {
|
|
dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
|
|
goto err_free_txport;
|
|
}
|
|
|
|
txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
|
|
|
|
if (vep.nr_of_link_frequencies != 1) {
|
|
ret = -EINVAL;
|
|
goto err_free_vep;
|
|
}
|
|
|
|
priv->tx_link_freq[0] = vep.link_frequencies[0];
|
|
priv->tx_data_rate = priv->tx_link_freq[0] * 2;
|
|
|
|
if (priv->tx_data_rate != MHZ(1600) &&
|
|
priv->tx_data_rate != MHZ(1200) &&
|
|
priv->tx_data_rate != MHZ(800) &&
|
|
priv->tx_data_rate != MHZ(400)) {
|
|
dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
|
|
ret = -EINVAL;
|
|
goto err_free_vep;
|
|
}
|
|
|
|
v4l2_fwnode_endpoint_free(&vep);
|
|
|
|
priv->txports[nport] = txport;
|
|
|
|
return 0;
|
|
|
|
err_free_vep:
|
|
v4l2_fwnode_endpoint_free(&vep);
|
|
err_free_txport:
|
|
kfree(txport);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 csi_tx_isr;
|
|
int ret;
|
|
|
|
ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
|
|
if (ret)
|
|
return;
|
|
|
|
if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
|
|
dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
|
|
|
|
if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
|
|
dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* RX ports
|
|
*/
|
|
|
|
static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
ret = regulator_enable(rxport->vpoc);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_disable_vpocs:
|
|
while (nport--) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
regulator_disable(rxport->vpoc);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
regulator_disable(rxport->vpoc);
|
|
}
|
|
}
|
|
|
|
static void ub960_rxport_clear_errors(struct ub960_data *priv,
|
|
unsigned int nport)
|
|
{
|
|
u8 v;
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
|
|
}
|
|
|
|
static void ub960_clear_rx_errors(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
|
|
ub960_rxport_clear_errors(priv, nport);
|
|
}
|
|
|
|
static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
|
|
unsigned int nport, s8 *strobe_pos)
|
|
{
|
|
u8 v;
|
|
u8 clk_delay, data_delay;
|
|
int ret;
|
|
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
|
|
|
|
clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
|
|
0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
|
|
|
|
data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
|
|
0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
|
|
|
|
*strobe_pos = data_delay - clk_delay;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
|
|
unsigned int nport, s8 strobe_pos)
|
|
{
|
|
u8 clk_delay, data_delay;
|
|
|
|
clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
|
|
data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
|
|
|
|
if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
|
|
clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
|
|
data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
else if (strobe_pos < 0)
|
|
clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
|
|
else if (strobe_pos > 0)
|
|
data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
|
|
}
|
|
|
|
static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
|
|
s8 strobe_min, s8 strobe_max)
|
|
{
|
|
/* Convert the signed strobe pos to positive zero based value */
|
|
strobe_min -= UB960_MIN_AEQ_STROBE_POS;
|
|
strobe_max -= UB960_MIN_AEQ_STROBE_POS;
|
|
|
|
ub960_write(priv, UB960_XR_SFILTER_CFG,
|
|
((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
|
|
((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
|
|
}
|
|
|
|
static int ub960_rxport_get_eq_level(struct ub960_data *priv,
|
|
unsigned int nport, u8 *eq_level)
|
|
{
|
|
int ret;
|
|
u8 v;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
*eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
|
|
(v & UB960_RR_AEQ_STATUS_STATUS_2);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_set_eq_level(struct ub960_data *priv,
|
|
unsigned int nport, u8 eq_level)
|
|
{
|
|
u8 eq_stage_1_select_value, eq_stage_2_select_value;
|
|
const unsigned int eq_stage_max = 7;
|
|
u8 v;
|
|
|
|
if (eq_level <= eq_stage_max) {
|
|
eq_stage_1_select_value = eq_level;
|
|
eq_stage_2_select_value = 0;
|
|
} else {
|
|
eq_stage_1_select_value = eq_stage_max;
|
|
eq_stage_2_select_value = eq_level - eq_stage_max;
|
|
}
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
|
|
|
|
v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
|
|
UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
|
|
v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
|
|
v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
|
|
v |= UB960_RR_AEQ_BYPASS_ENABLE;
|
|
|
|
ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
|
|
}
|
|
|
|
static void ub960_rxport_set_eq_range(struct ub960_data *priv,
|
|
unsigned int nport, u8 eq_min, u8 eq_max)
|
|
{
|
|
ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
|
|
(eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
|
|
(eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
|
|
|
|
/* Enable AEQ min setting */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
|
|
UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
|
|
UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
|
|
}
|
|
|
|
static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
/* We also set common settings here. Should be moved elsewhere. */
|
|
|
|
if (priv->strobe.manual) {
|
|
/* Disable AEQ_SFILTER_EN */
|
|
ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
|
|
UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
|
|
} else {
|
|
/* Enable SFILTER and error control */
|
|
ub960_write(priv, UB960_XR_AEQ_CTL1,
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
|
|
UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
|
|
|
|
/* Set AEQ strobe range */
|
|
ub960_rxport_set_strobe_range(priv, priv->strobe.min,
|
|
priv->strobe.max);
|
|
}
|
|
|
|
/* The rest are port specific */
|
|
|
|
if (priv->strobe.manual)
|
|
ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
|
|
else
|
|
ub960_rxport_set_strobe_pos(priv, nport, 0);
|
|
|
|
if (rxport->eq.manual_eq) {
|
|
ub960_rxport_set_eq_level(priv, nport,
|
|
rxport->eq.manual.eq_level);
|
|
|
|
/* Enable AEQ Bypass */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
|
|
UB960_RR_AEQ_BYPASS_ENABLE,
|
|
UB960_RR_AEQ_BYPASS_ENABLE);
|
|
} else {
|
|
ub960_rxport_set_eq_range(priv, nport,
|
|
rxport->eq.aeq.eq_level_min,
|
|
rxport->eq.aeq.eq_level_max);
|
|
|
|
/* Disable AEQ Bypass */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
|
|
UB960_RR_AEQ_BYPASS_ENABLE, 0);
|
|
}
|
|
}
|
|
|
|
static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
|
|
bool *ok)
|
|
{
|
|
u8 rx_port_sts1, rx_port_sts2;
|
|
u16 parity_errors;
|
|
u8 csi_rx_sts;
|
|
u8 csi_err_cnt;
|
|
u8 bcc_sts;
|
|
int ret;
|
|
bool errors;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
|
|
&rx_port_sts1);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
|
|
*ok = false;
|
|
return 0;
|
|
}
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
|
|
&rx_port_sts2);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
|
|
&csi_err_cnt);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
|
|
&parity_errors);
|
|
if (ret)
|
|
return ret;
|
|
|
|
errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
|
|
(bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
|
|
(csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
|
|
parity_errors;
|
|
|
|
*ok = !errors;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Wait for the RX ports to lock, have no errors and have stable strobe position
|
|
* and EQ level.
|
|
*/
|
|
static int ub960_rxport_wait_locks(struct ub960_data *priv,
|
|
unsigned long port_mask,
|
|
unsigned int *lock_mask)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned long timeout;
|
|
unsigned int link_ok_mask;
|
|
unsigned int missing;
|
|
unsigned int loops;
|
|
u8 nport;
|
|
int ret;
|
|
|
|
if (port_mask == 0) {
|
|
if (lock_mask)
|
|
*lock_mask = 0;
|
|
return 0;
|
|
}
|
|
|
|
if (port_mask >= BIT(priv->hw_data->num_rxports))
|
|
return -EINVAL;
|
|
|
|
timeout = jiffies + msecs_to_jiffies(1000);
|
|
loops = 0;
|
|
link_ok_mask = 0;
|
|
|
|
while (time_before(jiffies, timeout)) {
|
|
missing = 0;
|
|
|
|
for_each_set_bit(nport, &port_mask,
|
|
priv->hw_data->num_rxports) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
bool ok;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = ub960_rxport_link_ok(priv, nport, &ok);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* We want the link to be ok for two consecutive loops,
|
|
* as a link could get established just before our test
|
|
* and drop soon after.
|
|
*/
|
|
if (!ok || !(link_ok_mask & BIT(nport)))
|
|
missing++;
|
|
|
|
if (ok)
|
|
link_ok_mask |= BIT(nport);
|
|
else
|
|
link_ok_mask &= ~BIT(nport);
|
|
}
|
|
|
|
loops++;
|
|
|
|
if (missing == 0)
|
|
break;
|
|
|
|
msleep(50);
|
|
}
|
|
|
|
if (lock_mask)
|
|
*lock_mask = link_ok_mask;
|
|
|
|
dev_dbg(dev, "Wait locks done in %u loops\n", loops);
|
|
for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
s8 strobe_pos, eq_level;
|
|
u16 v;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
if (!(link_ok_mask & BIT(nport))) {
|
|
dev_dbg(dev, "\trx%u: not locked\n", nport);
|
|
continue;
|
|
}
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
|
|
|
|
ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
|
|
if (ret)
|
|
return ret;
|
|
|
|
dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
|
|
nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int mult;
|
|
unsigned int div;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
mult = 1;
|
|
div = 10;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
mult = 2;
|
|
div = 1;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
mult = 2;
|
|
div = 5;
|
|
break;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
|
|
return clk_get_rate(priv->refclk) * mult / div;
|
|
}
|
|
|
|
static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
return 2359400;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
return 47187500;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
return 9437500;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
struct device *dev = &priv->client->dev;
|
|
struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
|
|
struct i2c_board_info ser_info = {
|
|
.of_node = to_of_node(rxport->ser.fwnode),
|
|
.fwnode = rxport->ser.fwnode,
|
|
.platform_data = ser_pdata,
|
|
};
|
|
|
|
ser_pdata->port = nport;
|
|
ser_pdata->atr = priv->atr;
|
|
if (priv->hw_data->is_ub9702)
|
|
ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
|
|
else
|
|
ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
|
|
|
|
/*
|
|
* The serializer is added under the same i2c adapter as the
|
|
* deserializer. This is not quite right, as the serializer is behind
|
|
* the FPD-Link.
|
|
*/
|
|
ser_info.addr = rxport->ser.alias;
|
|
rxport->ser.client =
|
|
i2c_new_client_device(priv->client->adapter, &ser_info);
|
|
if (IS_ERR(rxport->ser.client)) {
|
|
dev_err(dev, "rx%u: cannot add %s i2c device", nport,
|
|
ser_info.type);
|
|
return PTR_ERR(rxport->ser.client);
|
|
}
|
|
|
|
dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
|
|
nport, rxport->ser.client->addr,
|
|
rxport->ser.client->adapter->nr, rxport->ser.client->addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
i2c_unregister_device(rxport->ser.client);
|
|
rxport->ser.client = NULL;
|
|
}
|
|
|
|
/* Add serializer i2c devices for all initialized ports */
|
|
static int ub960_rxport_add_serializers(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = ub960_rxport_add_serializer(priv, nport);
|
|
if (ret)
|
|
goto err_remove_sers;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_remove_sers:
|
|
while (nport--) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ub960_rxport_remove_serializer(priv, nport);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_rxport_remove_serializers(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ub960_rxport_remove_serializer(priv, nport);
|
|
}
|
|
}
|
|
|
|
static void ub960_init_tx_port(struct ub960_data *priv,
|
|
struct ub960_txport *txport)
|
|
{
|
|
unsigned int nport = txport->nport;
|
|
u8 csi_ctl = 0;
|
|
|
|
/*
|
|
* From the datasheet: "initial CSI Skew-Calibration
|
|
* sequence [...] should be set when operating at 1.6 Gbps"
|
|
*/
|
|
if (priv->tx_data_rate == MHZ(1600))
|
|
csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
|
|
|
|
csi_ctl |= (4 - txport->num_data_lanes) << 4;
|
|
|
|
ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
|
|
}
|
|
|
|
static int ub960_init_tx_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
u8 speed_select;
|
|
u8 pll_div;
|
|
|
|
/* TX ports */
|
|
|
|
switch (priv->tx_data_rate) {
|
|
case MHZ(1600):
|
|
default:
|
|
speed_select = 0;
|
|
pll_div = 0x10;
|
|
break;
|
|
case MHZ(1200):
|
|
speed_select = 1;
|
|
break;
|
|
case MHZ(800):
|
|
speed_select = 2;
|
|
pll_div = 0x10;
|
|
break;
|
|
case MHZ(400):
|
|
speed_select = 3;
|
|
pll_div = 0x10;
|
|
break;
|
|
}
|
|
|
|
ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
|
|
|
|
if (priv->hw_data->is_ub9702) {
|
|
ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
|
|
|
|
switch (priv->tx_data_rate) {
|
|
case MHZ(1600):
|
|
default:
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
|
|
break;
|
|
case MHZ(800):
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
|
|
break;
|
|
case MHZ(400):
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
if (!txport)
|
|
continue;
|
|
|
|
ub960_init_tx_port(priv, txport);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub960(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u32 bc_freq_val;
|
|
|
|
/*
|
|
* Back channel frequency select.
|
|
* Override FREQ_SELECT from the strap.
|
|
* 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
|
|
* 2 - 10 Mbps
|
|
* 6 - 50 Mbps (DS90UB953-Q1)
|
|
*
|
|
* Note that changing this setting will result in some errors on the back
|
|
* channel for a short period of time.
|
|
*/
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
|
|
bc_freq_val);
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
/* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
|
|
UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
|
|
0x3);
|
|
|
|
/*
|
|
* RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
|
|
*/
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
|
|
UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
|
|
0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
return;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
/* CSI-2 Mode (DS90UB953-Q1 compatible) */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
|
|
0x0);
|
|
|
|
break;
|
|
}
|
|
|
|
/* LV_POLARITY & FV_POLARITY */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
|
|
rxport->lv_fv_pol);
|
|
|
|
/* Enable all interrupt sources from this port */
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
|
|
|
|
/* Enable I2C_PASS_THROUGH */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
|
|
|
|
/* Enable I2C communication to the serializer via the alias addr */
|
|
ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
|
|
rxport->ser.alias << 1);
|
|
|
|
/* Configure EQ related settings */
|
|
ub960_rxport_config_eq(priv, nport);
|
|
|
|
/* Enable RX port */
|
|
ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u8 bc_freq_val;
|
|
u8 fpd_func_mode;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 5;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 4;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 6;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
fpd_func_mode = 2;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
fpd_func_mode = 2;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
|
|
bc_freq_val);
|
|
ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
|
|
|
|
/* set serdes_eq_mode = 1 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
|
|
|
|
/* enable serdes driver */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
|
|
|
|
/* set serdes_eq_offset=4 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
|
|
|
|
/* init default serdes_eq_max in 0xa9 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
|
|
|
|
/* init serdes_eq_min in 0xaa */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
|
|
|
|
/* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
|
|
ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
|
|
BIT(3), BIT(3));
|
|
|
|
/* RX port to half-rate */
|
|
ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
|
|
BIT(nport * 2));
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
bool first_time_power_up = true;
|
|
|
|
if (first_time_power_up) {
|
|
u8 v;
|
|
|
|
/* AEQ init */
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
|
|
}
|
|
|
|
/* enable serdes_eq_ctl2 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
|
|
|
|
/* enable serdes_eq_ctl1 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
|
|
|
|
/* enable serdes_eq_en */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
|
|
|
|
/* disable serdes_eq_override */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
|
|
|
|
/* disable serdes_gain_override */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u8 bc_freq_val;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
|
|
bc_freq_val);
|
|
|
|
/* FPD4 Sync Mode */
|
|
ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
|
|
|
|
/* add serdes_eq_offset of 4 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
|
|
|
|
/* FPD4 serdes_start_eq in 0x27: assign default */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
|
|
/* FPD4 serdes_end_eq in 0x28: assign default */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
|
|
|
|
/* set serdes_driver_mode into FPD IV mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
|
|
/* set FPD PBC drv into FPD IV mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
|
|
|
|
/* set serdes_system_init to 0x2f */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
|
|
/* set serdes_system_rst in reset mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
|
|
|
|
/* RX port to 7.55G mode */
|
|
ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
|
|
0 << (nport * 2));
|
|
|
|
ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
|
|
if (rxport->cdr_mode == RXPORT_CDR_FPD3)
|
|
ub960_init_rx_port_ub9702_fpd3(priv, rxport);
|
|
else /* RXPORT_CDR_FPD4 */
|
|
ub960_init_rx_port_ub9702_fpd4(priv, rxport);
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
/*
|
|
* RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
|
|
* 0b10 : 8-bit processing using upper 8 bits
|
|
*/
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
|
|
0x3 << 6, 0x2 << 6);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
return;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
|
|
break;
|
|
}
|
|
|
|
/* LV_POLARITY & FV_POLARITY */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
|
|
rxport->lv_fv_pol);
|
|
|
|
/* Enable all interrupt sources from this port */
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
|
|
|
|
/* Enable I2C_PASS_THROUGH */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
|
|
|
|
/* Enable I2C communication to the serializer via the alias addr */
|
|
ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
|
|
rxport->ser.alias << 1);
|
|
|
|
/* Enable RX port */
|
|
ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
|
|
|
|
if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
|
|
/* unreset 960 AEQ */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
|
|
}
|
|
}
|
|
|
|
static int ub960_init_rx_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
if (priv->hw_data->is_ub9702)
|
|
ub960_init_rx_port_ub9702(priv, rxport);
|
|
else
|
|
ub960_init_rx_port_ub960(priv, rxport);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 rx_port_sts1;
|
|
u8 rx_port_sts2;
|
|
u8 csi_rx_sts;
|
|
u8 bcc_sts;
|
|
int ret = 0;
|
|
|
|
/* Read interrupts (also clears most of them) */
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
|
|
&rx_port_sts1);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
|
|
&rx_port_sts2);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
|
|
&csi_rx_sts);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
|
|
&bcc_sts);
|
|
|
|
if (ret)
|
|
return;
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
|
|
&v);
|
|
if (!ret)
|
|
dev_err(dev, "rx%u parity errors: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
|
|
dev_err(dev, "rx%u BCC CRC error\n", nport);
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
|
|
dev_err(dev, "rx%u BCC SEQ error\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
|
|
dev_err(dev, "rx%u line length unstable\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
|
|
dev_err(dev, "rx%u FPD3 encode error\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
|
|
dev_err(dev, "rx%u buffer error\n", nport);
|
|
|
|
if (csi_rx_sts)
|
|
dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
|
|
dev_err(dev, "rx%u CSI ECC1 error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
|
|
dev_err(dev, "rx%u CSI ECC2 error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
|
|
dev_err(dev, "rx%u CSI checksum error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
|
|
dev_err(dev, "rx%u CSI length error\n", nport);
|
|
|
|
if (bcc_sts)
|
|
dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
|
|
dev_err(dev, "rx%u BCC response error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
|
|
dev_err(dev, "rx%u BCC slave timeout", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
|
|
dev_err(dev, "rx%u BCC slave error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
|
|
dev_err(dev, "rx%u BCC master timeout", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
|
|
dev_err(dev, "rx%u BCC master error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
|
|
dev_err(dev, "rx%u BCC sequence error", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
|
|
if (!ret)
|
|
dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
|
|
&v);
|
|
if (!ret)
|
|
dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
|
|
dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
|
|
(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
|
|
"locked" :
|
|
"unlocked",
|
|
(rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
|
|
"passed" :
|
|
"not passed",
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
|
|
"no clock" :
|
|
"clock ok",
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
|
|
"stable freq" :
|
|
"unstable freq");
|
|
}
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* V4L2
|
|
*/
|
|
|
|
/*
|
|
* The current implementation only supports a simple VC mapping, where all VCs
|
|
* from a one RX port will be mapped to the same VC. Also, the hardware
|
|
* dictates that all streams from an RX port must go to a single TX port.
|
|
*
|
|
* This function decides the target VC numbers for each RX port with a simple
|
|
* algorithm, so that for each TX port, we get VC numbers starting from 0,
|
|
* and counting up.
|
|
*
|
|
* E.g. if all four RX ports are in use, of which the first two go to the
|
|
* first TX port and the secont two go to the second TX port, we would get
|
|
* the following VCs for the four RX ports: 0, 1, 0, 1.
|
|
*
|
|
* TODO: implement a more sophisticated VC mapping. As the driver cannot know
|
|
* what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
|
|
* probably needs to be somehow configurable. Device tree?
|
|
*/
|
|
static void ub960_get_vc_maps(struct ub960_data *priv,
|
|
struct v4l2_subdev_state *state, u8 *vc)
|
|
{
|
|
u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
u8 handled_mask = 0;
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
unsigned int rx, tx;
|
|
|
|
rx = ub960_pad_to_port(priv, route->sink_pad);
|
|
if (BIT(rx) & handled_mask)
|
|
continue;
|
|
|
|
tx = ub960_pad_to_port(priv, route->source_pad);
|
|
|
|
vc[rx] = cur_vc[tx]++;
|
|
handled_mask |= BIT(rx);
|
|
}
|
|
}
|
|
|
|
static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "enable TX port %u\n", nport);
|
|
|
|
return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE);
|
|
}
|
|
|
|
static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "disable TX port %u\n", nport);
|
|
|
|
ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE, 0);
|
|
}
|
|
|
|
static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "enable RX port %u\n", nport);
|
|
|
|
/* Enable forwarding */
|
|
return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
|
|
}
|
|
|
|
static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "disable RX port %u\n", nport);
|
|
|
|
/* Disable forwarding */
|
|
ub960_update_bits(priv, UB960_SR_FWD_CTL1,
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport),
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport));
|
|
}
|
|
|
|
/*
|
|
* The driver only supports using a single VC for each source. This function
|
|
* checks that each source only provides streams using a single VC.
|
|
*/
|
|
static int ub960_validate_stream_vcs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
struct v4l2_mbus_frame_desc desc;
|
|
int ret;
|
|
u8 vc;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
|
|
rxport->source.pad, &desc);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
|
|
continue;
|
|
|
|
if (desc.num_entries == 0)
|
|
continue;
|
|
|
|
vc = desc.entry[0].bus.csi2.vc;
|
|
|
|
for (i = 1; i < desc.num_entries; i++) {
|
|
if (vc == desc.entry[i].bus.csi2.vc)
|
|
continue;
|
|
|
|
dev_err(&priv->client->dev,
|
|
"rx%u: source with multiple virtual-channels is not supported\n",
|
|
nport);
|
|
return -ENODEV;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
|
|
struct v4l2_subdev_state *state)
|
|
{
|
|
u8 fwd_ctl;
|
|
struct {
|
|
u32 num_streams;
|
|
u8 pixel_dt;
|
|
u8 meta_dt;
|
|
u32 meta_lines;
|
|
u32 tx_port;
|
|
} rx_data[UB960_MAX_RX_NPORTS] = {};
|
|
u8 vc_map[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
ret = ub960_validate_stream_vcs(priv);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ub960_get_vc_maps(priv, state, vc_map);
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
struct ub960_rxport *rxport;
|
|
struct ub960_txport *txport;
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
const struct ub960_format_info *ub960_fmt;
|
|
unsigned int nport;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
rxport = priv->rxports[nport];
|
|
if (!rxport)
|
|
return -EINVAL;
|
|
|
|
txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
|
|
if (!txport)
|
|
return -EINVAL;
|
|
|
|
rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
|
|
|
|
rx_data[nport].num_streams++;
|
|
|
|
/* For the rest, we are only interested in parallel busses */
|
|
if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
|
|
rxport->rx_mode == RXPORT_MODE_CSI2_ASYNC)
|
|
continue;
|
|
|
|
if (rx_data[nport].num_streams > 2)
|
|
return -EPIPE;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state,
|
|
route->sink_pad,
|
|
route->sink_stream);
|
|
if (!fmt)
|
|
return -EPIPE;
|
|
|
|
ub960_fmt = ub960_find_format(fmt->code);
|
|
if (!ub960_fmt)
|
|
return -EPIPE;
|
|
|
|
if (ub960_fmt->meta) {
|
|
if (fmt->height > 3) {
|
|
dev_err(&priv->client->dev,
|
|
"rx%u: unsupported metadata height %u\n",
|
|
nport, fmt->height);
|
|
return -EPIPE;
|
|
}
|
|
|
|
rx_data[nport].meta_dt = ub960_fmt->datatype;
|
|
rx_data[nport].meta_lines = fmt->height;
|
|
} else {
|
|
rx_data[nport].pixel_dt = ub960_fmt->datatype;
|
|
}
|
|
}
|
|
|
|
/* Configure RX ports */
|
|
|
|
/*
|
|
* Keep all port forwardings disabled by default. Forwarding will be
|
|
* enabled in ub960_enable_rx_port.
|
|
*/
|
|
fwd_ctl = GENMASK(7, 4);
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
u8 vc = vc_map[nport];
|
|
|
|
if (rx_data[nport].num_streams == 0)
|
|
continue;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
|
|
rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
|
|
|
|
ub960_rxport_write(priv, rxport->nport,
|
|
UB960_RR_RAW_EMBED_DTYPE,
|
|
(rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
|
|
rx_data[nport].meta_dt);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
if (!priv->hw_data->is_ub9702) {
|
|
/* Map all VCs from this port to the same VC */
|
|
ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
|
|
} else {
|
|
unsigned int i;
|
|
|
|
/* Map all VCs from this port to VC(nport) */
|
|
for (i = 0; i < 8; i++)
|
|
ub960_rxport_write(priv, nport,
|
|
UB960_RR_VC_ID_MAP(i),
|
|
nport);
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
if (rx_data[nport].tx_port == 1)
|
|
fwd_ctl |= BIT(nport); /* forward to TX1 */
|
|
else
|
|
fwd_ctl &= ~BIT(nport); /* forward to TX0 */
|
|
}
|
|
|
|
ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_update_streaming_status(struct ub960_data *priv)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < UB960_MAX_NPORTS; i++) {
|
|
if (priv->stream_enable_mask[i])
|
|
break;
|
|
}
|
|
|
|
priv->streaming = i < UB960_MAX_NPORTS;
|
|
}
|
|
|
|
static int ub960_enable_streams(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state, u32 source_pad,
|
|
u64 source_streams_mask)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int failed_port;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
if (!priv->streaming) {
|
|
dev_dbg(dev, "Prepare for streaming\n");
|
|
ret = ub960_configure_ports_for_streaming(priv, state);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/* Enable TX port if not yet enabled */
|
|
if (!priv->stream_enable_mask[source_pad]) {
|
|
ret = ub960_enable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
priv->stream_enable_mask[source_pad] |= source_streams_mask;
|
|
|
|
/* Collect sink streams per pad which we need to enable */
|
|
for_each_active_route(&state->routing, route) {
|
|
if (route->source_pad != source_pad)
|
|
continue;
|
|
|
|
if (!(source_streams_mask & BIT_ULL(route->source_stream)))
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
sink_streams[nport] |= BIT_ULL(route->sink_stream);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
/* Enable the RX port if not yet enabled */
|
|
if (!priv->stream_enable_mask[nport]) {
|
|
ret = ub960_enable_rx_port(priv, nport);
|
|
if (ret) {
|
|
failed_port = nport;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
priv->stream_enable_mask[nport] |= sink_streams[nport];
|
|
|
|
dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_enable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret) {
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
|
|
failed_port = nport;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
priv->streaming = true;
|
|
|
|
return 0;
|
|
|
|
err:
|
|
for (nport = 0; nport < failed_port; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_disable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret)
|
|
dev_err(dev, "Failed to disable streams: %d\n", ret);
|
|
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
/* Disable RX port if no active streams */
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
}
|
|
|
|
priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
|
|
|
|
if (!priv->stream_enable_mask[source_pad])
|
|
ub960_disable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
|
|
ub960_update_streaming_status(priv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_disable_streams(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
u32 source_pad, u64 source_streams_mask)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
/* Collect sink streams per pad which we need to disable */
|
|
for_each_active_route(&state->routing, route) {
|
|
if (route->source_pad != source_pad)
|
|
continue;
|
|
|
|
if (!(source_streams_mask & BIT_ULL(route->source_stream)))
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
sink_streams[nport] |= BIT_ULL(route->sink_stream);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_disable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret)
|
|
dev_err(dev, "Failed to disable streams: %d\n", ret);
|
|
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
/* Disable RX port if no active streams */
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
}
|
|
|
|
/* Disable TX port if no active streams */
|
|
|
|
priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
|
|
|
|
if (!priv->stream_enable_mask[source_pad])
|
|
ub960_disable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
|
|
ub960_update_streaming_status(priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int _ub960_set_routing(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
struct v4l2_subdev_krouting *routing)
|
|
{
|
|
static const struct v4l2_mbus_framefmt format = {
|
|
.width = 640,
|
|
.height = 480,
|
|
.code = MEDIA_BUS_FMT_UYVY8_1X16,
|
|
.field = V4L2_FIELD_NONE,
|
|
.colorspace = V4L2_COLORSPACE_SRGB,
|
|
.ycbcr_enc = V4L2_YCBCR_ENC_601,
|
|
.quantization = V4L2_QUANTIZATION_LIM_RANGE,
|
|
.xfer_func = V4L2_XFER_FUNC_SRGB,
|
|
};
|
|
int ret;
|
|
|
|
/*
|
|
* Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
|
|
* frame desc is made dynamically allocated.
|
|
*/
|
|
|
|
if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
|
|
return -E2BIG;
|
|
|
|
ret = v4l2_subdev_routing_validate(sd, routing,
|
|
V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
|
|
V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_set_routing(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
enum v4l2_subdev_format_whence which,
|
|
struct v4l2_subdev_krouting *routing)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
|
|
return -EBUSY;
|
|
|
|
return _ub960_set_routing(sd, state, routing);
|
|
}
|
|
|
|
static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
|
|
struct v4l2_mbus_frame_desc *fd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct v4l2_subdev_route *route;
|
|
struct v4l2_subdev_state *state;
|
|
int ret = 0;
|
|
struct device *dev = &priv->client->dev;
|
|
u8 vc_map[UB960_MAX_RX_NPORTS] = {};
|
|
|
|
if (!ub960_pad_is_source(priv, pad))
|
|
return -EINVAL;
|
|
|
|
memset(fd, 0, sizeof(*fd));
|
|
|
|
fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
|
|
|
|
state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
|
|
|
|
ub960_get_vc_maps(priv, state, vc_map);
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
|
|
struct v4l2_mbus_frame_desc source_fd;
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
|
|
if (route->source_pad != pad)
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
|
|
get_frame_desc,
|
|
priv->rxports[nport]->source.pad,
|
|
&source_fd);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Failed to get source frame desc for pad %u\n",
|
|
route->sink_pad);
|
|
goto out_unlock;
|
|
}
|
|
|
|
for (i = 0; i < source_fd.num_entries; i++) {
|
|
if (source_fd.entry[i].stream == route->sink_stream) {
|
|
source_entry = &source_fd.entry[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!source_entry) {
|
|
dev_err(dev,
|
|
"Failed to find stream from source frame desc\n");
|
|
ret = -EPIPE;
|
|
goto out_unlock;
|
|
}
|
|
|
|
fd->entry[fd->num_entries].stream = route->source_stream;
|
|
fd->entry[fd->num_entries].flags = source_entry->flags;
|
|
fd->entry[fd->num_entries].length = source_entry->length;
|
|
fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
|
|
|
|
fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
|
|
|
|
if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
|
|
fd->entry[fd->num_entries].bus.csi2.dt =
|
|
source_entry->bus.csi2.dt;
|
|
} else {
|
|
const struct ub960_format_info *ub960_fmt;
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state, pad,
|
|
route->source_stream);
|
|
|
|
if (!fmt) {
|
|
ret = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
ub960_fmt = ub960_find_format(fmt->code);
|
|
if (!ub960_fmt) {
|
|
dev_err(dev, "Unable to find format\n");
|
|
ret = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
fd->entry[fd->num_entries].bus.csi2.dt =
|
|
ub960_fmt->datatype;
|
|
}
|
|
|
|
fd->num_entries++;
|
|
}
|
|
|
|
out_unlock:
|
|
v4l2_subdev_unlock_state(state);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_set_fmt(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
struct v4l2_subdev_format *format)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
|
|
if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
|
|
return -EBUSY;
|
|
|
|
/* No transcoding, source and sink formats must match. */
|
|
if (ub960_pad_is_source(priv, format->pad))
|
|
return v4l2_subdev_get_fmt(sd, state, format);
|
|
|
|
/*
|
|
* Default to the first format if the requested media bus code isn't
|
|
* supported.
|
|
*/
|
|
if (!ub960_find_format(format->format.code))
|
|
format->format.code = ub960_formats[0].code;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state, format->pad,
|
|
format->stream);
|
|
if (!fmt)
|
|
return -EINVAL;
|
|
|
|
*fmt = format->format;
|
|
|
|
fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
|
|
format->stream);
|
|
if (!fmt)
|
|
return -EINVAL;
|
|
|
|
*fmt = format->format;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_init_cfg(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
struct v4l2_subdev_route routes[] = {
|
|
{
|
|
.sink_pad = 0,
|
|
.sink_stream = 0,
|
|
.source_pad = priv->hw_data->num_rxports,
|
|
.source_stream = 0,
|
|
.flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
|
|
},
|
|
};
|
|
|
|
struct v4l2_subdev_krouting routing = {
|
|
.num_routes = ARRAY_SIZE(routes),
|
|
.routes = routes,
|
|
};
|
|
|
|
return _ub960_set_routing(sd, state, &routing);
|
|
}
|
|
|
|
static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
|
|
.enable_streams = ub960_enable_streams,
|
|
.disable_streams = ub960_disable_streams,
|
|
|
|
.set_routing = ub960_set_routing,
|
|
.get_frame_desc = ub960_get_frame_desc,
|
|
|
|
.get_fmt = v4l2_subdev_get_fmt,
|
|
.set_fmt = ub960_set_fmt,
|
|
|
|
.init_cfg = ub960_init_cfg,
|
|
};
|
|
|
|
static int ub960_log_status(struct v4l2_subdev *sd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_subdev_state *state;
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
u16 v16 = 0;
|
|
u8 v = 0;
|
|
u8 id[UB960_SR_FPD3_RX_ID_LEN];
|
|
|
|
state = v4l2_subdev_lock_and_get_active_state(sd);
|
|
|
|
for (i = 0; i < sizeof(id); i++)
|
|
ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
|
|
|
|
dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
dev_info(dev, "TX %u\n", nport);
|
|
|
|
if (!txport) {
|
|
dev_info(dev, "\tNot initialized\n");
|
|
continue;
|
|
}
|
|
|
|
ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
|
|
dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
|
|
v & (u8)BIT(0));
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tframe counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tframe error counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tline counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tline error counter %u\n", v16);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
u8 eq_level;
|
|
s8 strobe_pos;
|
|
unsigned int i;
|
|
|
|
dev_info(dev, "RX %u\n", nport);
|
|
|
|
if (!rxport) {
|
|
dev_info(dev, "\tNot initialized\n");
|
|
continue;
|
|
}
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
|
|
|
|
if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
|
|
dev_info(dev, "\tLocked\n");
|
|
else
|
|
dev_info(dev, "\tNot locked\n");
|
|
|
|
dev_info(dev, "\trx_port_sts1 %#02x\n", v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
|
|
dev_info(dev, "\trx_port_sts2 %#02x\n", v);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
|
|
dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
|
|
dev_info(dev, "\tparity errors %u\n", v16);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
|
|
dev_info(dev, "\tlines per frame %u\n", v16);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
|
|
dev_info(dev, "\tbytes per line %u\n", v16);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
|
|
dev_info(dev, "\tcsi_err_counter %u\n", v);
|
|
|
|
/* Strobe */
|
|
|
|
ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
|
|
|
|
dev_info(dev, "\t%s strobe\n",
|
|
(v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
|
|
"Manual");
|
|
|
|
if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
|
|
ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
|
|
|
|
dev_info(dev, "\tStrobe range [%d, %d]\n",
|
|
((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
|
|
((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
|
|
}
|
|
|
|
ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
|
|
|
|
dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
|
|
|
|
/* EQ */
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
|
|
|
|
dev_info(dev, "\t%s EQ\n",
|
|
(v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
|
|
"Adaptive");
|
|
|
|
if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
|
|
|
|
dev_info(dev, "\tEQ range [%u, %u]\n",
|
|
(v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
|
|
(v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
|
|
}
|
|
|
|
if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
|
|
dev_info(dev, "\tEQ level %u\n", eq_level);
|
|
|
|
/* GPIOs */
|
|
for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
|
|
u8 ctl_reg;
|
|
u8 ctl_shift;
|
|
|
|
ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
|
|
ctl_shift = (i % 2) * 4;
|
|
|
|
ub960_rxport_read(priv, nport, ctl_reg, &v);
|
|
|
|
dev_info(dev, "\tGPIO%u: mode %u\n", i,
|
|
(v >> ctl_shift) & 0xf);
|
|
}
|
|
}
|
|
|
|
v4l2_subdev_unlock_state(state);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
|
|
.log_status = ub960_log_status,
|
|
.subscribe_event = v4l2_ctrl_subdev_subscribe_event,
|
|
.unsubscribe_event = v4l2_event_subdev_unsubscribe,
|
|
};
|
|
|
|
static const struct v4l2_subdev_ops ub960_subdev_ops = {
|
|
.core = &ub960_subdev_core_ops,
|
|
.pad = &ub960_pad_ops,
|
|
};
|
|
|
|
static const struct media_entity_operations ub960_entity_ops = {
|
|
.get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
|
|
.link_validate = v4l2_subdev_link_validate,
|
|
.has_pad_interdep = v4l2_subdev_has_pad_interdep,
|
|
};
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* Core
|
|
*/
|
|
|
|
static irqreturn_t ub960_handle_events(int irq, void *arg)
|
|
{
|
|
struct ub960_data *priv = arg;
|
|
unsigned int i;
|
|
u8 int_sts;
|
|
u8 fwd_sts;
|
|
int ret;
|
|
|
|
ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
|
|
if (ret || !int_sts)
|
|
return IRQ_NONE;
|
|
|
|
dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
|
|
|
|
ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
|
|
if (ret)
|
|
return IRQ_NONE;
|
|
|
|
dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
|
|
|
|
for (i = 0; i < priv->hw_data->num_txports; i++) {
|
|
if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
|
|
ub960_csi_handle_events(priv, i);
|
|
}
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
if (!priv->rxports[i])
|
|
continue;
|
|
|
|
if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
|
|
ub960_rxport_handle_events(priv, i);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void ub960_handler_work(struct work_struct *work)
|
|
{
|
|
struct delayed_work *dwork = to_delayed_work(work);
|
|
struct ub960_data *priv =
|
|
container_of(dwork, struct ub960_data, poll_work);
|
|
|
|
ub960_handle_events(0, priv);
|
|
|
|
schedule_delayed_work(&priv->poll_work,
|
|
msecs_to_jiffies(UB960_POLL_TIME_MS));
|
|
}
|
|
|
|
static void ub960_txport_free_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
if (!txport)
|
|
continue;
|
|
|
|
kfree(txport);
|
|
priv->txports[nport] = NULL;
|
|
}
|
|
}
|
|
|
|
static void ub960_rxport_free_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
fwnode_handle_put(rxport->source.ep_fwnode);
|
|
fwnode_handle_put(rxport->ser.fwnode);
|
|
|
|
kfree(rxport);
|
|
priv->rxports[nport] = NULL;
|
|
}
|
|
}
|
|
|
|
static int
|
|
ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
|
|
struct fwnode_handle *link_fwnode,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int nport = rxport->nport;
|
|
u32 rx_mode;
|
|
u32 cdr_mode;
|
|
s32 strobe_pos;
|
|
u32 eq_level;
|
|
u32 ser_i2c_alias;
|
|
int ret;
|
|
|
|
cdr_mode = RXPORT_CDR_FPD3;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
|
|
if (ret < 0 && ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,cdr-mode", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (cdr_mode > RXPORT_CDR_LAST) {
|
|
dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
|
|
dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
|
|
return -EINVAL;
|
|
}
|
|
|
|
rxport->cdr_mode = cdr_mode;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
|
|
if (ret < 0) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,rx-mode", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (rx_mode > RXPORT_MODE_LAST) {
|
|
dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
|
|
return -EINVAL;
|
|
}
|
|
|
|
switch (rx_mode) {
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
|
|
rx_mode);
|
|
return -EINVAL;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
rxport->rx_mode = rx_mode;
|
|
|
|
/* EQ & Strobe related */
|
|
|
|
/* Defaults */
|
|
rxport->eq.manual_eq = false;
|
|
rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
|
|
rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
|
|
&strobe_pos);
|
|
if (ret) {
|
|
if (ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,strobe-pos", ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
|
|
strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
|
|
dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
|
|
nport, strobe_pos);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* NOTE: ignored unless global manual strobe pos is also set */
|
|
rxport->eq.strobe_pos = strobe_pos;
|
|
if (!priv->strobe.manual)
|
|
dev_warn(dev,
|
|
"rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
|
|
nport);
|
|
}
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
|
|
if (ret) {
|
|
if (ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,eq-level", ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
if (eq_level > UB960_MAX_EQ_LEVEL) {
|
|
dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
|
|
nport, eq_level);
|
|
return -EINVAL;
|
|
}
|
|
|
|
rxport->eq.manual_eq = true;
|
|
rxport->eq.manual.eq_level = eq_level;
|
|
}
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
|
|
&ser_i2c_alias);
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"i2c-alias", ret);
|
|
return ret;
|
|
}
|
|
rxport->ser.alias = ser_i2c_alias;
|
|
|
|
rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
|
|
if (!rxport->ser.fwnode) {
|
|
dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
|
|
struct fwnode_handle *ep_fwnode,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_fwnode_endpoint vep = {};
|
|
unsigned int nport = rxport->nport;
|
|
bool hsync_hi;
|
|
bool vsync_hi;
|
|
int ret;
|
|
|
|
rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
|
|
if (!rxport->source.ep_fwnode) {
|
|
dev_err(dev, "rx%u: no remote endpoint\n", nport);
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* We currently have properties only for RAW modes */
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
break;
|
|
default:
|
|
return 0;
|
|
}
|
|
|
|
vep.bus_type = V4L2_MBUS_PARALLEL;
|
|
ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
|
|
goto err_put_source_ep_fwnode;
|
|
}
|
|
|
|
hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
|
|
vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
|
|
|
|
/* LineValid and FrameValid are inverse to the h/vsync active */
|
|
rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
|
|
(vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
|
|
|
|
return 0;
|
|
|
|
err_put_source_ep_fwnode:
|
|
fwnode_handle_put(rxport->source.ep_fwnode);
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
|
|
struct fwnode_handle *link_fwnode,
|
|
struct fwnode_handle *ep_fwnode)
|
|
{
|
|
static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
|
|
"vpoc0", "vpoc1", "vpoc2", "vpoc3"
|
|
};
|
|
struct device *dev = &priv->client->dev;
|
|
struct ub960_rxport *rxport;
|
|
int ret;
|
|
|
|
rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
|
|
if (!rxport)
|
|
return -ENOMEM;
|
|
|
|
priv->rxports[nport] = rxport;
|
|
|
|
rxport->nport = nport;
|
|
rxport->priv = priv;
|
|
|
|
ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
|
|
if (ret)
|
|
goto err_free_rxport;
|
|
|
|
rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
|
|
if (IS_ERR(rxport->vpoc)) {
|
|
ret = PTR_ERR(rxport->vpoc);
|
|
if (ret == -ENODEV) {
|
|
rxport->vpoc = NULL;
|
|
} else {
|
|
dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
|
|
nport, ret);
|
|
goto err_put_remote_fwnode;
|
|
}
|
|
}
|
|
|
|
ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
|
|
if (ret)
|
|
goto err_put_remote_fwnode;
|
|
|
|
return 0;
|
|
|
|
err_put_remote_fwnode:
|
|
fwnode_handle_put(rxport->ser.fwnode);
|
|
err_free_rxport:
|
|
priv->rxports[nport] = NULL;
|
|
kfree(rxport);
|
|
return ret;
|
|
}
|
|
|
|
static struct fwnode_handle *
|
|
ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
|
|
unsigned int nport)
|
|
{
|
|
struct fwnode_handle *link_fwnode;
|
|
int ret;
|
|
|
|
fwnode_for_each_child_node(links_fwnode, link_fwnode) {
|
|
u32 link_num;
|
|
|
|
if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
|
|
continue;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
|
|
if (ret) {
|
|
fwnode_handle_put(link_fwnode);
|
|
return NULL;
|
|
}
|
|
|
|
if (nport == link_num)
|
|
return link_fwnode;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxports(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct fwnode_handle *links_fwnode;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
|
|
if (!links_fwnode) {
|
|
dev_err(dev, "'links' node missing\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* Defaults, recommended by TI */
|
|
priv->strobe.min = 2;
|
|
priv->strobe.max = 3;
|
|
|
|
priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct fwnode_handle *link_fwnode;
|
|
struct fwnode_handle *ep_fwnode;
|
|
|
|
link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
|
|
if (!link_fwnode)
|
|
continue;
|
|
|
|
ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
|
|
nport, 0, 0);
|
|
if (!ep_fwnode) {
|
|
fwnode_handle_put(link_fwnode);
|
|
continue;
|
|
}
|
|
|
|
ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
|
|
ep_fwnode);
|
|
|
|
fwnode_handle_put(link_fwnode);
|
|
fwnode_handle_put(ep_fwnode);
|
|
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to parse RX port\n", nport);
|
|
goto err_put_links;
|
|
}
|
|
}
|
|
|
|
fwnode_handle_put(links_fwnode);
|
|
|
|
return 0;
|
|
|
|
err_put_links:
|
|
fwnode_handle_put(links_fwnode);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_parse_dt_txports(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u32 nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
unsigned int port = nport + priv->hw_data->num_rxports;
|
|
struct fwnode_handle *ep_fwnode;
|
|
|
|
ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
|
|
port, 0, 0);
|
|
if (!ep_fwnode)
|
|
continue;
|
|
|
|
ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
|
|
|
|
fwnode_handle_put(ep_fwnode);
|
|
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_parse_dt(struct ub960_data *priv)
|
|
{
|
|
int ret;
|
|
|
|
ret = ub960_parse_dt_rxports(priv);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_parse_dt_txports(priv);
|
|
if (ret)
|
|
goto err_free_rxports;
|
|
|
|
return 0;
|
|
|
|
err_free_rxports:
|
|
ub960_rxport_free_ports(priv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(notifier->sd);
|
|
struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
|
|
struct device *dev = &priv->client->dev;
|
|
u8 nport = rxport->nport;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
ret = media_entity_get_fwnode_pad(&subdev->entity,
|
|
rxport->source.ep_fwnode,
|
|
MEDIA_PAD_FL_SOURCE);
|
|
if (ret < 0) {
|
|
dev_err(dev, "Failed to find pad for %s\n", subdev->name);
|
|
return ret;
|
|
}
|
|
|
|
rxport->source.sd = subdev;
|
|
rxport->source.pad = ret;
|
|
|
|
ret = media_create_pad_link(&rxport->source.sd->entity,
|
|
rxport->source.pad, &priv->sd.entity, nport,
|
|
MEDIA_LNK_FL_ENABLED |
|
|
MEDIA_LNK_FL_IMMUTABLE);
|
|
if (ret) {
|
|
dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
|
|
rxport->source.sd->name, rxport->source.pad,
|
|
priv->sd.name, nport);
|
|
return ret;
|
|
}
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
|
|
dev_dbg(dev, "Waiting for more subdevs to be bound\n");
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
|
|
|
|
rxport->source.sd = NULL;
|
|
}
|
|
|
|
static const struct v4l2_async_notifier_operations ub960_notify_ops = {
|
|
.bound = ub960_notify_bound,
|
|
.unbind = ub960_notify_unbind,
|
|
};
|
|
|
|
static int ub960_v4l2_notifier_register(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
v4l2_async_nf_init(&priv->notifier);
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
struct ub960_rxport *rxport = priv->rxports[i];
|
|
struct ub960_asd *asd;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
asd = v4l2_async_nf_add_fwnode(&priv->notifier,
|
|
rxport->source.ep_fwnode,
|
|
struct ub960_asd);
|
|
if (IS_ERR(asd)) {
|
|
dev_err(dev, "Failed to add subdev for source %u: %pe",
|
|
i, asd);
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
return PTR_ERR(asd);
|
|
}
|
|
|
|
asd->rxport = rxport;
|
|
}
|
|
|
|
priv->notifier.ops = &ub960_notify_ops;
|
|
|
|
ret = v4l2_async_subdev_nf_register(&priv->sd, &priv->notifier);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to register subdev_notifier");
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
|
|
{
|
|
v4l2_async_nf_unregister(&priv->notifier);
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
}
|
|
|
|
static int ub960_create_subdev(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
|
|
|
|
v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
|
|
priv->sd.ctrl_handler = &priv->ctrl_handler;
|
|
|
|
v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
|
|
ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
|
|
priv->tx_link_freq);
|
|
|
|
if (priv->ctrl_handler.error) {
|
|
ret = priv->ctrl_handler.error;
|
|
goto err_free_ctrl;
|
|
}
|
|
|
|
priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
|
|
V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
|
|
priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
|
|
priv->sd.entity.ops = &ub960_entity_ops;
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
|
|
priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
|
|
MEDIA_PAD_FL_SINK :
|
|
MEDIA_PAD_FL_SOURCE;
|
|
}
|
|
|
|
ret = media_entity_pads_init(&priv->sd.entity,
|
|
priv->hw_data->num_rxports +
|
|
priv->hw_data->num_txports,
|
|
priv->pads);
|
|
if (ret)
|
|
goto err_free_ctrl;
|
|
|
|
priv->sd.state_lock = priv->sd.ctrl_handler->lock;
|
|
|
|
ret = v4l2_subdev_init_finalize(&priv->sd);
|
|
if (ret)
|
|
goto err_entity_cleanup;
|
|
|
|
ret = ub960_v4l2_notifier_register(priv);
|
|
if (ret) {
|
|
dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
|
|
goto err_subdev_cleanup;
|
|
}
|
|
|
|
ret = v4l2_async_register_subdev(&priv->sd);
|
|
if (ret) {
|
|
dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
|
|
goto err_unreg_notif;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_unreg_notif:
|
|
ub960_v4l2_notifier_unregister(priv);
|
|
err_subdev_cleanup:
|
|
v4l2_subdev_cleanup(&priv->sd);
|
|
err_entity_cleanup:
|
|
media_entity_cleanup(&priv->sd.entity);
|
|
err_free_ctrl:
|
|
v4l2_ctrl_handler_free(&priv->ctrl_handler);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_destroy_subdev(struct ub960_data *priv)
|
|
{
|
|
ub960_v4l2_notifier_unregister(priv);
|
|
v4l2_async_unregister_subdev(&priv->sd);
|
|
|
|
v4l2_subdev_cleanup(&priv->sd);
|
|
|
|
media_entity_cleanup(&priv->sd.entity);
|
|
v4l2_ctrl_handler_free(&priv->ctrl_handler);
|
|
}
|
|
|
|
static const struct regmap_config ub960_regmap_config = {
|
|
.name = "ds90ub960",
|
|
|
|
.reg_bits = 8,
|
|
.val_bits = 8,
|
|
|
|
.max_register = 0xff,
|
|
|
|
/*
|
|
* We do locking in the driver to cover the TX/RX port selection and the
|
|
* indirect register access.
|
|
*/
|
|
.disable_locking = true,
|
|
};
|
|
|
|
static void ub960_reset(struct ub960_data *priv, bool reset_regs)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
u8 bit;
|
|
|
|
bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
|
|
UB960_SR_RESET_DIGITAL_RESET0;
|
|
|
|
ub960_write(priv, UB960_SR_RESET, bit);
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
|
|
(v & bit) == 0, 2000, 100000);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
if (ret)
|
|
dev_err(dev, "reset failed: %d\n", ret);
|
|
}
|
|
|
|
static int ub960_get_hw_resources(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
|
|
if (IS_ERR(priv->regmap))
|
|
return PTR_ERR(priv->regmap);
|
|
|
|
priv->vddio = devm_regulator_get(dev, "vddio");
|
|
if (IS_ERR(priv->vddio))
|
|
return dev_err_probe(dev, PTR_ERR(priv->vddio),
|
|
"cannot get VDDIO regulator\n");
|
|
|
|
/* get power-down pin from DT */
|
|
priv->pd_gpio =
|
|
devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
|
|
if (IS_ERR(priv->pd_gpio))
|
|
return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
|
|
"Cannot get powerdown GPIO\n");
|
|
|
|
priv->refclk = devm_clk_get(dev, "refclk");
|
|
if (IS_ERR(priv->refclk))
|
|
return dev_err_probe(dev, PTR_ERR(priv->refclk),
|
|
"Cannot get REFCLK\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_enable_core_hw(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 rev_mask;
|
|
int ret;
|
|
u8 dev_sts;
|
|
u8 refclk_freq;
|
|
|
|
ret = regulator_enable(priv->vddio);
|
|
if (ret)
|
|
return dev_err_probe(dev, ret,
|
|
"failed to enable VDDIO regulator\n");
|
|
|
|
ret = clk_prepare_enable(priv->refclk);
|
|
if (ret) {
|
|
dev_err_probe(dev, ret, "Failed to enable refclk\n");
|
|
goto err_disable_vddio;
|
|
}
|
|
|
|
if (priv->pd_gpio) {
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
/* wait min 2 ms for reset to complete */
|
|
fsleep(2000);
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 0);
|
|
/* wait min 2 ms for power up to finish */
|
|
fsleep(2000);
|
|
}
|
|
|
|
ub960_reset(priv, true);
|
|
|
|
/* Runtime check register accessibility */
|
|
ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
|
|
if (ret) {
|
|
dev_err_probe(dev, ret, "Cannot read first register, abort\n");
|
|
goto err_pd_gpio;
|
|
}
|
|
|
|
dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
|
|
rev_mask);
|
|
|
|
ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
|
|
!!(dev_sts & BIT(4)), refclk_freq,
|
|
clk_get_rate(priv->refclk) / 1000000);
|
|
|
|
/* Disable all RX ports by default */
|
|
ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
/* release GPIO lock */
|
|
if (priv->hw_data->is_ub9702) {
|
|
ret = ub960_update_bits(priv, UB960_SR_RESET,
|
|
UB960_SR_RESET_GPIO_LOCK_RELEASE,
|
|
UB960_SR_RESET_GPIO_LOCK_RELEASE);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_pd_gpio:
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
clk_disable_unprepare(priv->refclk);
|
|
err_disable_vddio:
|
|
regulator_disable(priv->vddio);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_disable_core_hw(struct ub960_data *priv)
|
|
{
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
clk_disable_unprepare(priv->refclk);
|
|
regulator_disable(priv->vddio);
|
|
}
|
|
|
|
static int ub960_probe(struct i2c_client *client)
|
|
{
|
|
struct device *dev = &client->dev;
|
|
struct ub960_data *priv;
|
|
unsigned int port_lock_mask;
|
|
unsigned int port_mask;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
|
|
if (!priv)
|
|
return -ENOMEM;
|
|
|
|
priv->client = client;
|
|
|
|
priv->hw_data = device_get_match_data(dev);
|
|
|
|
mutex_init(&priv->reg_lock);
|
|
|
|
INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
|
|
|
|
/*
|
|
* Initialize these to invalid values so that the first reg writes will
|
|
* configure the target.
|
|
*/
|
|
priv->reg_current.indirect_target = 0xff;
|
|
priv->reg_current.rxport = 0xff;
|
|
priv->reg_current.txport = 0xff;
|
|
|
|
ret = ub960_get_hw_resources(priv);
|
|
if (ret)
|
|
goto err_mutex_destroy;
|
|
|
|
ret = ub960_enable_core_hw(priv);
|
|
if (ret)
|
|
goto err_mutex_destroy;
|
|
|
|
ret = ub960_parse_dt(priv);
|
|
if (ret)
|
|
goto err_disable_core_hw;
|
|
|
|
ret = ub960_init_tx_ports(priv);
|
|
if (ret)
|
|
goto err_free_ports;
|
|
|
|
ret = ub960_rxport_enable_vpocs(priv);
|
|
if (ret)
|
|
goto err_free_ports;
|
|
|
|
ret = ub960_init_rx_ports(priv);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
ub960_reset(priv, false);
|
|
|
|
port_mask = 0;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
port_mask |= BIT(nport);
|
|
}
|
|
|
|
ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
if (port_mask != port_lock_mask) {
|
|
ret = -EIO;
|
|
dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
|
|
goto err_disable_vpocs;
|
|
}
|
|
|
|
/*
|
|
* Clear any errors caused by switching the RX port settings while
|
|
* probing.
|
|
*/
|
|
ub960_clear_rx_errors(priv);
|
|
|
|
ret = ub960_init_atr(priv);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
ret = ub960_rxport_add_serializers(priv);
|
|
if (ret)
|
|
goto err_uninit_atr;
|
|
|
|
ret = ub960_create_subdev(priv);
|
|
if (ret)
|
|
goto err_free_sers;
|
|
|
|
if (client->irq)
|
|
dev_warn(dev, "irq support not implemented, using polling\n");
|
|
|
|
schedule_delayed_work(&priv->poll_work,
|
|
msecs_to_jiffies(UB960_POLL_TIME_MS));
|
|
|
|
return 0;
|
|
|
|
err_free_sers:
|
|
ub960_rxport_remove_serializers(priv);
|
|
err_uninit_atr:
|
|
ub960_uninit_atr(priv);
|
|
err_disable_vpocs:
|
|
ub960_rxport_disable_vpocs(priv);
|
|
err_free_ports:
|
|
ub960_rxport_free_ports(priv);
|
|
ub960_txport_free_ports(priv);
|
|
err_disable_core_hw:
|
|
ub960_disable_core_hw(priv);
|
|
err_mutex_destroy:
|
|
mutex_destroy(&priv->reg_lock);
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_remove(struct i2c_client *client)
|
|
{
|
|
struct v4l2_subdev *sd = i2c_get_clientdata(client);
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
cancel_delayed_work_sync(&priv->poll_work);
|
|
|
|
ub960_destroy_subdev(priv);
|
|
ub960_rxport_remove_serializers(priv);
|
|
ub960_uninit_atr(priv);
|
|
ub960_rxport_disable_vpocs(priv);
|
|
ub960_rxport_free_ports(priv);
|
|
ub960_txport_free_ports(priv);
|
|
ub960_disable_core_hw(priv);
|
|
mutex_destroy(&priv->reg_lock);
|
|
}
|
|
|
|
static const struct ub960_hw_data ds90ub960_hw = {
|
|
.model = "ub960",
|
|
.num_rxports = 4,
|
|
.num_txports = 2,
|
|
};
|
|
|
|
static const struct ub960_hw_data ds90ub9702_hw = {
|
|
.model = "ub9702",
|
|
.num_rxports = 4,
|
|
.num_txports = 2,
|
|
.is_ub9702 = true,
|
|
.is_fpdlink4 = true,
|
|
};
|
|
|
|
static const struct i2c_device_id ub960_id[] = {
|
|
{ "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
|
|
{ "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(i2c, ub960_id);
|
|
|
|
static const struct of_device_id ub960_dt_ids[] = {
|
|
{ .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
|
|
{ .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(of, ub960_dt_ids);
|
|
|
|
static struct i2c_driver ds90ub960_driver = {
|
|
.probe = ub960_probe,
|
|
.remove = ub960_remove,
|
|
.id_table = ub960_id,
|
|
.driver = {
|
|
.name = "ds90ub960",
|
|
.of_match_table = ub960_dt_ids,
|
|
},
|
|
};
|
|
module_i2c_driver(ds90ub960_driver);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
|
|
MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
|
|
MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
|
|
MODULE_IMPORT_NS(I2C_ATR);
|