mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-15 06:55:13 +08:00
05428f66fc
Allow using FPD-Link in async mode. The driver handles it correctly, but the mode was blocked at probe time as there wasn't HW to test this with. Now the mode has been tested, and it works. Signed-off-by: Tomi Valkeinen <tomi.valkeinen@ideasonboard.com> Reviewed-by: Andy Shevchenko <andriy.shevchenko@linux.intel.com> Reviewed-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com> Signed-off-by: Sakari Ailus <sakari.ailus@linux.intel.com> Signed-off-by: Hans Verkuil <hverkuil-cisco@xs4all.nl>
4059 lines
100 KiB
C
4059 lines
100 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Driver for the Texas Instruments DS90UB960-Q1 video deserializer
|
|
*
|
|
* Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
|
|
* Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
|
|
*/
|
|
|
|
/*
|
|
* (Possible) TODOs:
|
|
*
|
|
* - PM for serializer and remote peripherals. We need to manage:
|
|
* - VPOC
|
|
* - Power domain? Regulator? Somehow any remote device should be able to
|
|
* cause the VPOC to be turned on.
|
|
* - Link between the deserializer and the serializer
|
|
* - Related to VPOC management. We probably always want to turn on the VPOC
|
|
* and then enable the link.
|
|
* - Serializer's services: i2c, gpios, power
|
|
* - The serializer needs to resume before the remote peripherals can
|
|
* e.g. use the i2c.
|
|
* - How to handle gpios? Reserving a gpio essentially keeps the provider
|
|
* (serializer) always powered on.
|
|
* - Do we need a new bus for the FPD-Link? At the moment the serializers
|
|
* are children of the same i2c-adapter where the deserializer resides.
|
|
* - i2c-atr could be made embeddable instead of allocatable.
|
|
*/
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/fwnode.h>
|
|
#include <linux/gpio/consumer.h>
|
|
#include <linux/i2c-atr.h>
|
|
#include <linux/i2c.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/property.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/regulator/consumer.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/workqueue.h>
|
|
|
|
#include <media/i2c/ds90ub9xx.h>
|
|
#include <media/mipi-csi2.h>
|
|
#include <media/v4l2-ctrls.h>
|
|
#include <media/v4l2-event.h>
|
|
#include <media/v4l2-fwnode.h>
|
|
#include <media/v4l2-subdev.h>
|
|
|
|
#define MHZ(v) ((u32)((v) * 1000000U))
|
|
|
|
#define UB960_POLL_TIME_MS 500
|
|
|
|
#define UB960_MAX_RX_NPORTS 4
|
|
#define UB960_MAX_TX_NPORTS 2
|
|
#define UB960_MAX_NPORTS (UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
|
|
|
|
#define UB960_MAX_PORT_ALIASES 8
|
|
|
|
#define UB960_NUM_BC_GPIOS 4
|
|
|
|
/*
|
|
* Register map
|
|
*
|
|
* 0x00-0x32 Shared (UB960_SR)
|
|
* 0x33-0x3a CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
|
|
* 0x4c Shared (UB960_SR)
|
|
* 0x4d-0x7f FPD-Link RX, per-port paged (UB960_RR)
|
|
* 0xb0-0xbf Shared (UB960_SR)
|
|
* 0xd0-0xdf FPD-Link RX, per-port paged (UB960_RR)
|
|
* 0xf0-0xf5 Shared (UB960_SR)
|
|
* 0xf8-0xfb Shared (UB960_SR)
|
|
* All others Reserved
|
|
*
|
|
* Register prefixes:
|
|
* UB960_SR_* = Shared register
|
|
* UB960_RR_* = FPD-Link RX, per-port paged register
|
|
* UB960_TR_* = CSI-2 TX, per-port paged register
|
|
* UB960_XR_* = Reserved register
|
|
* UB960_IR_* = Indirect register
|
|
*/
|
|
|
|
#define UB960_SR_I2C_DEV_ID 0x00
|
|
#define UB960_SR_RESET 0x01
|
|
#define UB960_SR_RESET_DIGITAL_RESET1 BIT(1)
|
|
#define UB960_SR_RESET_DIGITAL_RESET0 BIT(0)
|
|
#define UB960_SR_RESET_GPIO_LOCK_RELEASE BIT(5)
|
|
|
|
#define UB960_SR_GEN_CONFIG 0x02
|
|
#define UB960_SR_REV_MASK 0x03
|
|
#define UB960_SR_DEVICE_STS 0x04
|
|
#define UB960_SR_PAR_ERR_THOLD_HI 0x05
|
|
#define UB960_SR_PAR_ERR_THOLD_LO 0x06
|
|
#define UB960_SR_BCC_WDOG_CTL 0x07
|
|
#define UB960_SR_I2C_CTL1 0x08
|
|
#define UB960_SR_I2C_CTL2 0x09
|
|
#define UB960_SR_SCL_HIGH_TIME 0x0a
|
|
#define UB960_SR_SCL_LOW_TIME 0x0b
|
|
#define UB960_SR_RX_PORT_CTL 0x0c
|
|
#define UB960_SR_IO_CTL 0x0d
|
|
#define UB960_SR_GPIO_PIN_STS 0x0e
|
|
#define UB960_SR_GPIO_INPUT_CTL 0x0f
|
|
#define UB960_SR_GPIO_PIN_CTL(n) (0x10 + (n)) /* n < UB960_NUM_GPIOS */
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL 5
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT 2
|
|
#define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN BIT(0)
|
|
|
|
#define UB960_SR_FS_CTL 0x18
|
|
#define UB960_SR_FS_HIGH_TIME_1 0x19
|
|
#define UB960_SR_FS_HIGH_TIME_0 0x1a
|
|
#define UB960_SR_FS_LOW_TIME_1 0x1b
|
|
#define UB960_SR_FS_LOW_TIME_0 0x1c
|
|
#define UB960_SR_MAX_FRM_HI 0x1d
|
|
#define UB960_SR_MAX_FRM_LO 0x1e
|
|
#define UB960_SR_CSI_PLL_CTL 0x1f
|
|
|
|
#define UB960_SR_FWD_CTL1 0x20
|
|
#define UB960_SR_FWD_CTL1_PORT_DIS(n) BIT((n) + 4)
|
|
|
|
#define UB960_SR_FWD_CTL2 0x21
|
|
#define UB960_SR_FWD_STS 0x22
|
|
|
|
#define UB960_SR_INTERRUPT_CTL 0x23
|
|
#define UB960_SR_INTERRUPT_CTL_INT_EN BIT(7)
|
|
#define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0 BIT(4)
|
|
#define UB960_SR_INTERRUPT_CTL_IE_RX(n) BIT((n)) /* rxport[n] IRQ */
|
|
|
|
#define UB960_SR_INTERRUPT_STS 0x24
|
|
#define UB960_SR_INTERRUPT_STS_INT BIT(7)
|
|
#define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n) BIT(4 + (n)) /* txport[n] IRQ */
|
|
#define UB960_SR_INTERRUPT_STS_IS_RX(n) BIT((n)) /* rxport[n] IRQ */
|
|
|
|
#define UB960_SR_TS_CONFIG 0x25
|
|
#define UB960_SR_TS_CONTROL 0x26
|
|
#define UB960_SR_TS_LINE_HI 0x27
|
|
#define UB960_SR_TS_LINE_LO 0x28
|
|
#define UB960_SR_TS_STATUS 0x29
|
|
#define UB960_SR_TIMESTAMP_P0_HI 0x2a
|
|
#define UB960_SR_TIMESTAMP_P0_LO 0x2b
|
|
#define UB960_SR_TIMESTAMP_P1_HI 0x2c
|
|
#define UB960_SR_TIMESTAMP_P1_LO 0x2d
|
|
|
|
#define UB960_SR_CSI_PORT_SEL 0x32
|
|
|
|
#define UB960_TR_CSI_CTL 0x33
|
|
#define UB960_TR_CSI_CTL_CSI_CAL_EN BIT(6)
|
|
#define UB960_TR_CSI_CTL_CSI_CONTS_CLOCK BIT(1)
|
|
#define UB960_TR_CSI_CTL_CSI_ENABLE BIT(0)
|
|
|
|
#define UB960_TR_CSI_CTL2 0x34
|
|
#define UB960_TR_CSI_STS 0x35
|
|
#define UB960_TR_CSI_TX_ICR 0x36
|
|
|
|
#define UB960_TR_CSI_TX_ISR 0x37
|
|
#define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR BIT(3)
|
|
#define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR BIT(1)
|
|
|
|
#define UB960_TR_CSI_TEST_CTL 0x38
|
|
#define UB960_TR_CSI_TEST_PATT_HI 0x39
|
|
#define UB960_TR_CSI_TEST_PATT_LO 0x3a
|
|
|
|
#define UB960_XR_SFILTER_CFG 0x41
|
|
#define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT 4
|
|
#define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT 0
|
|
|
|
#define UB960_XR_AEQ_CTL1 0x42
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK BIT(6)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING BIT(5)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY BIT(4)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK \
|
|
(UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK | \
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
|
|
#define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN BIT(0)
|
|
|
|
#define UB960_XR_AEQ_ERR_THOLD 0x43
|
|
|
|
#define UB960_RR_BCC_ERR_CTL 0x46
|
|
#define UB960_RR_BCC_STATUS 0x47
|
|
#define UB960_RR_BCC_STATUS_SEQ_ERROR BIT(5)
|
|
#define UB960_RR_BCC_STATUS_MASTER_ERR BIT(4)
|
|
#define UB960_RR_BCC_STATUS_MASTER_TO BIT(3)
|
|
#define UB960_RR_BCC_STATUS_SLAVE_ERR BIT(2)
|
|
#define UB960_RR_BCC_STATUS_SLAVE_TO BIT(1)
|
|
#define UB960_RR_BCC_STATUS_RESP_ERR BIT(0)
|
|
#define UB960_RR_BCC_STATUS_ERROR_MASK \
|
|
(UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
|
|
UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR | \
|
|
UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
|
|
|
|
#define UB960_RR_FPD3_CAP 0x4a
|
|
#define UB960_RR_RAW_EMBED_DTYPE 0x4b
|
|
#define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT 6
|
|
|
|
#define UB960_SR_FPD3_PORT_SEL 0x4c
|
|
|
|
#define UB960_RR_RX_PORT_STS1 0x4d
|
|
#define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR BIT(5)
|
|
#define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG BIT(4)
|
|
#define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR BIT(3)
|
|
#define UB960_RR_RX_PORT_STS1_PARITY_ERROR BIT(2)
|
|
#define UB960_RR_RX_PORT_STS1_PORT_PASS BIT(1)
|
|
#define UB960_RR_RX_PORT_STS1_LOCK_STS BIT(0)
|
|
#define UB960_RR_RX_PORT_STS1_ERROR_MASK \
|
|
(UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
|
|
UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
|
|
UB960_RR_RX_PORT_STS1_PARITY_ERROR)
|
|
|
|
#define UB960_RR_RX_PORT_STS2 0x4e
|
|
#define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE BIT(7)
|
|
#define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG BIT(6)
|
|
#define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR BIT(5)
|
|
#define UB960_RR_RX_PORT_STS2_BUFFER_ERROR BIT(4)
|
|
#define UB960_RR_RX_PORT_STS2_CSI_ERROR BIT(3)
|
|
#define UB960_RR_RX_PORT_STS2_FREQ_STABLE BIT(2)
|
|
#define UB960_RR_RX_PORT_STS2_CABLE_FAULT BIT(1)
|
|
#define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG BIT(0)
|
|
#define UB960_RR_RX_PORT_STS2_ERROR_MASK \
|
|
UB960_RR_RX_PORT_STS2_BUFFER_ERROR
|
|
|
|
#define UB960_RR_RX_FREQ_HIGH 0x4f
|
|
#define UB960_RR_RX_FREQ_LOW 0x50
|
|
#define UB960_RR_SENSOR_STS_0 0x51
|
|
#define UB960_RR_SENSOR_STS_1 0x52
|
|
#define UB960_RR_SENSOR_STS_2 0x53
|
|
#define UB960_RR_SENSOR_STS_3 0x54
|
|
#define UB960_RR_RX_PAR_ERR_HI 0x55
|
|
#define UB960_RR_RX_PAR_ERR_LO 0x56
|
|
#define UB960_RR_BIST_ERR_COUNT 0x57
|
|
|
|
#define UB960_RR_BCC_CONFIG 0x58
|
|
#define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH BIT(6)
|
|
#define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK GENMASK(2, 0)
|
|
|
|
#define UB960_RR_DATAPATH_CTL1 0x59
|
|
#define UB960_RR_DATAPATH_CTL2 0x5a
|
|
#define UB960_RR_SER_ID 0x5b
|
|
#define UB960_RR_SER_ALIAS_ID 0x5c
|
|
|
|
/* For these two register sets: n < UB960_MAX_PORT_ALIASES */
|
|
#define UB960_RR_SLAVE_ID(n) (0x5d + (n))
|
|
#define UB960_RR_SLAVE_ALIAS(n) (0x65 + (n))
|
|
|
|
#define UB960_RR_PORT_CONFIG 0x6d
|
|
#define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK GENMASK(1, 0)
|
|
|
|
#define UB960_RR_BC_GPIO_CTL(n) (0x6e + (n)) /* n < 2 */
|
|
#define UB960_RR_RAW10_ID 0x70
|
|
#define UB960_RR_RAW10_ID_VC_SHIFT 6
|
|
#define UB960_RR_RAW10_ID_DT_SHIFT 0
|
|
|
|
#define UB960_RR_RAW12_ID 0x71
|
|
#define UB960_RR_CSI_VC_MAP 0x72
|
|
#define UB960_RR_CSI_VC_MAP_SHIFT(x) ((x) * 2)
|
|
|
|
#define UB960_RR_LINE_COUNT_HI 0x73
|
|
#define UB960_RR_LINE_COUNT_LO 0x74
|
|
#define UB960_RR_LINE_LEN_1 0x75
|
|
#define UB960_RR_LINE_LEN_0 0x76
|
|
#define UB960_RR_FREQ_DET_CTL 0x77
|
|
#define UB960_RR_MAILBOX_1 0x78
|
|
#define UB960_RR_MAILBOX_2 0x79
|
|
|
|
#define UB960_RR_CSI_RX_STS 0x7a
|
|
#define UB960_RR_CSI_RX_STS_LENGTH_ERR BIT(3)
|
|
#define UB960_RR_CSI_RX_STS_CKSUM_ERR BIT(2)
|
|
#define UB960_RR_CSI_RX_STS_ECC2_ERR BIT(1)
|
|
#define UB960_RR_CSI_RX_STS_ECC1_ERR BIT(0)
|
|
#define UB960_RR_CSI_RX_STS_ERROR_MASK \
|
|
(UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
|
|
UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
|
|
|
|
#define UB960_RR_CSI_ERR_COUNTER 0x7b
|
|
#define UB960_RR_PORT_CONFIG2 0x7c
|
|
#define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
|
|
#define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
|
|
|
|
#define UB960_RR_PORT_CONFIG2_LV_POL_LOW BIT(1)
|
|
#define UB960_RR_PORT_CONFIG2_FV_POL_LOW BIT(0)
|
|
|
|
#define UB960_RR_PORT_PASS_CTL 0x7d
|
|
#define UB960_RR_SEN_INT_RISE_CTL 0x7e
|
|
#define UB960_RR_SEN_INT_FALL_CTL 0x7f
|
|
|
|
#define UB960_SR_CSI_FRAME_COUNT_HI(n) (0x90 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_COUNT_LO(n) (0x91 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n) (0x92 + 8 * (n))
|
|
#define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n) (0x93 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_COUNT_HI(n) (0x94 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_COUNT_LO(n) (0x95 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_ERR_COUNT_HI(n) (0x96 + 8 * (n))
|
|
#define UB960_SR_CSI_LINE_ERR_COUNT_LO(n) (0x97 + 8 * (n))
|
|
|
|
#define UB960_XR_REFCLK_FREQ 0xa5 /* UB960 */
|
|
|
|
#define UB960_RR_VC_ID_MAP(x) (0xa0 + (x)) /* UB9702 */
|
|
|
|
#define UB960_SR_IND_ACC_CTL 0xb0
|
|
#define UB960_SR_IND_ACC_CTL_IA_AUTO_INC BIT(1)
|
|
|
|
#define UB960_SR_IND_ACC_ADDR 0xb1
|
|
#define UB960_SR_IND_ACC_DATA 0xb2
|
|
#define UB960_SR_BIST_CONTROL 0xb3
|
|
#define UB960_SR_MODE_IDX_STS 0xb8
|
|
#define UB960_SR_LINK_ERROR_COUNT 0xb9
|
|
#define UB960_SR_FPD3_ENC_CTL 0xba
|
|
#define UB960_SR_FV_MIN_TIME 0xbc
|
|
#define UB960_SR_GPIO_PD_CTL 0xbe
|
|
|
|
#define UB960_SR_FPD_RATE_CFG 0xc2 /* UB9702 */
|
|
#define UB960_SR_CSI_PLL_DIV 0xc9 /* UB9702 */
|
|
|
|
#define UB960_RR_PORT_DEBUG 0xd0
|
|
#define UB960_RR_AEQ_CTL2 0xd2
|
|
#define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR BIT(2)
|
|
|
|
#define UB960_RR_AEQ_STATUS 0xd3
|
|
#define UB960_RR_AEQ_STATUS_STATUS_2 GENMASK(5, 3)
|
|
#define UB960_RR_AEQ_STATUS_STATUS_1 GENMASK(2, 0)
|
|
|
|
#define UB960_RR_AEQ_BYPASS 0xd4
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT 5
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK GENMASK(7, 5)
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT 1
|
|
#define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK GENMASK(3, 1)
|
|
#define UB960_RR_AEQ_BYPASS_ENABLE BIT(0)
|
|
|
|
#define UB960_RR_AEQ_MIN_MAX 0xd5
|
|
#define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT 4
|
|
#define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT 0
|
|
|
|
#define UB960_RR_SFILTER_STS_0 0xd6
|
|
#define UB960_RR_SFILTER_STS_1 0xd7
|
|
#define UB960_RR_PORT_ICR_HI 0xd8
|
|
#define UB960_RR_PORT_ICR_LO 0xd9
|
|
#define UB960_RR_PORT_ISR_HI 0xda
|
|
#define UB960_RR_PORT_ISR_LO 0xdb
|
|
#define UB960_RR_FC_GPIO_STS 0xdc
|
|
#define UB960_RR_FC_GPIO_ICR 0xdd
|
|
#define UB960_RR_SEN_INT_RISE_STS 0xde
|
|
#define UB960_RR_SEN_INT_FALL_STS 0xdf
|
|
|
|
#define UB960_RR_CHANNEL_MODE 0xe4 /* UB9702 */
|
|
|
|
#define UB960_SR_FPD3_RX_ID(n) (0xf0 + (n))
|
|
#define UB960_SR_FPD3_RX_ID_LEN 6
|
|
|
|
#define UB960_SR_I2C_RX_ID(n) (0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
|
|
|
|
/* Indirect register blocks */
|
|
#define UB960_IND_TARGET_PAT_GEN 0x00
|
|
#define UB960_IND_TARGET_RX_ANA(n) (0x01 + (n))
|
|
#define UB960_IND_TARGET_CSI_CSIPLL_REG_1 0x92 /* UB9702 */
|
|
#define UB960_IND_TARGET_CSI_ANA 0x07
|
|
|
|
/* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
|
|
|
|
#define UB960_IR_PGEN_CTL 0x01
|
|
#define UB960_IR_PGEN_CTL_PGEN_ENABLE BIT(0)
|
|
|
|
#define UB960_IR_PGEN_CFG 0x02
|
|
#define UB960_IR_PGEN_CSI_DI 0x03
|
|
#define UB960_IR_PGEN_LINE_SIZE1 0x04
|
|
#define UB960_IR_PGEN_LINE_SIZE0 0x05
|
|
#define UB960_IR_PGEN_BAR_SIZE1 0x06
|
|
#define UB960_IR_PGEN_BAR_SIZE0 0x07
|
|
#define UB960_IR_PGEN_ACT_LPF1 0x08
|
|
#define UB960_IR_PGEN_ACT_LPF0 0x09
|
|
#define UB960_IR_PGEN_TOT_LPF1 0x0a
|
|
#define UB960_IR_PGEN_TOT_LPF0 0x0b
|
|
#define UB960_IR_PGEN_LINE_PD1 0x0c
|
|
#define UB960_IR_PGEN_LINE_PD0 0x0d
|
|
#define UB960_IR_PGEN_VBP 0x0e
|
|
#define UB960_IR_PGEN_VFP 0x0f
|
|
#define UB960_IR_PGEN_COLOR(n) (0x10 + (n)) /* n < 15 */
|
|
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK 0x08
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY BIT(3)
|
|
#define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK GENMASK(2, 0)
|
|
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA 0x09
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY BIT(3)
|
|
#define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK GENMASK(2, 0)
|
|
|
|
/* EQ related */
|
|
|
|
#define UB960_MIN_AEQ_STROBE_POS -7
|
|
#define UB960_MAX_AEQ_STROBE_POS 7
|
|
|
|
#define UB960_MANUAL_STROBE_EXTRA_DELAY 6
|
|
|
|
#define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
|
|
#define UB960_MAX_MANUAL_STROBE_POS (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
|
|
#define UB960_NUM_MANUAL_STROBE_POS (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
|
|
|
|
#define UB960_MIN_EQ_LEVEL 0
|
|
#define UB960_MAX_EQ_LEVEL 14
|
|
#define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
|
|
|
|
struct ub960_hw_data {
|
|
const char *model;
|
|
u8 num_rxports;
|
|
u8 num_txports;
|
|
bool is_ub9702;
|
|
bool is_fpdlink4;
|
|
};
|
|
|
|
enum ub960_rxport_mode {
|
|
RXPORT_MODE_RAW10 = 0,
|
|
RXPORT_MODE_RAW12_HF = 1,
|
|
RXPORT_MODE_RAW12_LF = 2,
|
|
RXPORT_MODE_CSI2_SYNC = 3,
|
|
RXPORT_MODE_CSI2_ASYNC = 4,
|
|
RXPORT_MODE_LAST = RXPORT_MODE_CSI2_ASYNC,
|
|
};
|
|
|
|
enum ub960_rxport_cdr {
|
|
RXPORT_CDR_FPD3 = 0,
|
|
RXPORT_CDR_FPD4 = 1,
|
|
RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
|
|
};
|
|
|
|
struct ub960_rxport {
|
|
struct ub960_data *priv;
|
|
u8 nport; /* RX port number, and index in priv->rxport[] */
|
|
|
|
struct {
|
|
struct v4l2_subdev *sd;
|
|
u16 pad;
|
|
struct fwnode_handle *ep_fwnode;
|
|
} source;
|
|
|
|
/* Serializer */
|
|
struct {
|
|
struct fwnode_handle *fwnode;
|
|
struct i2c_client *client;
|
|
unsigned short alias; /* I2C alias (lower 7 bits) */
|
|
struct ds90ub9xx_platform_data pdata;
|
|
} ser;
|
|
|
|
enum ub960_rxport_mode rx_mode;
|
|
enum ub960_rxport_cdr cdr_mode;
|
|
|
|
u8 lv_fv_pol; /* LV and FV polarities */
|
|
|
|
struct regulator *vpoc;
|
|
|
|
/* EQ settings */
|
|
struct {
|
|
bool manual_eq;
|
|
|
|
s8 strobe_pos;
|
|
|
|
union {
|
|
struct {
|
|
u8 eq_level_min;
|
|
u8 eq_level_max;
|
|
} aeq;
|
|
|
|
struct {
|
|
u8 eq_level;
|
|
} manual;
|
|
};
|
|
} eq;
|
|
|
|
const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
|
|
};
|
|
|
|
struct ub960_asd {
|
|
struct v4l2_async_connection base;
|
|
struct ub960_rxport *rxport;
|
|
};
|
|
|
|
static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_connection *asd)
|
|
{
|
|
return container_of(asd, struct ub960_asd, base);
|
|
}
|
|
|
|
struct ub960_txport {
|
|
struct ub960_data *priv;
|
|
u8 nport; /* TX port number, and index in priv->txport[] */
|
|
|
|
u32 num_data_lanes;
|
|
bool non_continous_clk;
|
|
};
|
|
|
|
struct ub960_data {
|
|
const struct ub960_hw_data *hw_data;
|
|
struct i2c_client *client; /* for shared local registers */
|
|
struct regmap *regmap;
|
|
|
|
/* lock for register access */
|
|
struct mutex reg_lock;
|
|
|
|
struct clk *refclk;
|
|
|
|
struct regulator *vddio;
|
|
|
|
struct gpio_desc *pd_gpio;
|
|
struct delayed_work poll_work;
|
|
struct ub960_rxport *rxports[UB960_MAX_RX_NPORTS];
|
|
struct ub960_txport *txports[UB960_MAX_TX_NPORTS];
|
|
|
|
struct v4l2_subdev sd;
|
|
struct media_pad pads[UB960_MAX_NPORTS];
|
|
|
|
struct v4l2_ctrl_handler ctrl_handler;
|
|
struct v4l2_async_notifier notifier;
|
|
|
|
u32 tx_data_rate; /* Nominal data rate (Gb/s) */
|
|
s64 tx_link_freq[1];
|
|
|
|
struct i2c_atr *atr;
|
|
|
|
struct {
|
|
u8 rxport;
|
|
u8 txport;
|
|
u8 indirect_target;
|
|
} reg_current;
|
|
|
|
bool streaming;
|
|
|
|
u8 stored_fwd_ctl;
|
|
|
|
u64 stream_enable_mask[UB960_MAX_NPORTS];
|
|
|
|
/* These are common to all ports */
|
|
struct {
|
|
bool manual;
|
|
|
|
s8 min;
|
|
s8 max;
|
|
} strobe;
|
|
};
|
|
|
|
static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
|
|
{
|
|
return container_of(sd, struct ub960_data, sd);
|
|
}
|
|
|
|
static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
|
|
{
|
|
return pad < priv->hw_data->num_rxports;
|
|
}
|
|
|
|
static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
|
|
{
|
|
return pad >= priv->hw_data->num_rxports;
|
|
}
|
|
|
|
static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
|
|
{
|
|
if (ub960_pad_is_sink(priv, pad))
|
|
return pad;
|
|
else
|
|
return pad - priv->hw_data->num_rxports;
|
|
}
|
|
|
|
struct ub960_format_info {
|
|
u32 code;
|
|
u32 bpp;
|
|
u8 datatype;
|
|
bool meta;
|
|
};
|
|
|
|
static const struct ub960_format_info ub960_formats[] = {
|
|
{ .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
{ .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
|
|
|
|
{ .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
{ .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
|
|
};
|
|
|
|
static const struct ub960_format_info *ub960_find_format(u32 code)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
|
|
if (ub960_formats[i].code == code)
|
|
return &ub960_formats[i];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* Basic device access
|
|
*/
|
|
|
|
static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
__be16 __v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = be16_to_cpu(__v);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.rxport == nport)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
|
|
(nport << 4) | BIT(nport));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
|
|
nport, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.rxport = nport;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u16 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
__be16 __v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_rxport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = be16_to_cpu(__v);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_select(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.txport == nport)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
|
|
(nport << 4) | BIT(nport));
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
|
|
nport, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.txport = nport;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_read(priv->regmap, reg, &v);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, reg, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_txport_select(priv, nport);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_update_bits(priv->regmap, reg, mask, val);
|
|
if (ret)
|
|
dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
|
|
__func__, reg, ret);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&priv->reg_lock);
|
|
|
|
if (priv->reg_current.indirect_target == block)
|
|
return 0;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
|
|
if (ret) {
|
|
dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
|
|
__func__, block, ret);
|
|
return ret;
|
|
}
|
|
|
|
priv->reg_current.indirect_target = block;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
*val = v;
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
|
|
u8 mask, u8 val)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
int ret;
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = ub960_select_ind_reg_block(priv, block);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
|
|
val);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
|
|
block, reg, ret);
|
|
goto out_unlock;
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* I2C-ATR (address translator)
|
|
*/
|
|
|
|
static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
|
|
const struct i2c_client *client, u16 alias)
|
|
{
|
|
struct ub960_data *priv = i2c_atr_get_driver_data(atr);
|
|
struct ub960_rxport *rxport = priv->rxports[chan_id];
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int reg_idx;
|
|
|
|
for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
|
|
if (!rxport->aliased_clients[reg_idx])
|
|
break;
|
|
}
|
|
|
|
if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
|
|
dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
|
|
return -EADDRNOTAVAIL;
|
|
}
|
|
|
|
rxport->aliased_clients[reg_idx] = client;
|
|
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
|
|
client->addr << 1);
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
|
|
alias << 1);
|
|
|
|
dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
|
|
rxport->nport, client->addr, alias, reg_idx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
|
|
const struct i2c_client *client)
|
|
{
|
|
struct ub960_data *priv = i2c_atr_get_driver_data(atr);
|
|
struct ub960_rxport *rxport = priv->rxports[chan_id];
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int reg_idx;
|
|
|
|
for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
|
|
if (rxport->aliased_clients[reg_idx] == client)
|
|
break;
|
|
}
|
|
|
|
if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
|
|
dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
|
|
rxport->nport, client->addr);
|
|
return;
|
|
}
|
|
|
|
rxport->aliased_clients[reg_idx] = NULL;
|
|
|
|
ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
|
|
|
|
dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
|
|
client->addr, reg_idx);
|
|
}
|
|
|
|
static const struct i2c_atr_ops ub960_atr_ops = {
|
|
.attach_client = ub960_atr_attach_client,
|
|
.detach_client = ub960_atr_detach_client,
|
|
};
|
|
|
|
static int ub960_init_atr(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct i2c_adapter *parent_adap = priv->client->adapter;
|
|
|
|
priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
|
|
priv->hw_data->num_rxports);
|
|
if (IS_ERR(priv->atr))
|
|
return PTR_ERR(priv->atr);
|
|
|
|
i2c_atr_set_driver_data(priv->atr, priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_uninit_atr(struct ub960_data *priv)
|
|
{
|
|
i2c_atr_delete(priv->atr);
|
|
priv->atr = NULL;
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* TX ports
|
|
*/
|
|
|
|
static int ub960_parse_dt_txport(struct ub960_data *priv,
|
|
struct fwnode_handle *ep_fwnode,
|
|
u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_fwnode_endpoint vep = {};
|
|
struct ub960_txport *txport;
|
|
int ret;
|
|
|
|
txport = kzalloc(sizeof(*txport), GFP_KERNEL);
|
|
if (!txport)
|
|
return -ENOMEM;
|
|
|
|
txport->priv = priv;
|
|
txport->nport = nport;
|
|
|
|
vep.bus_type = V4L2_MBUS_CSI2_DPHY;
|
|
ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
|
|
if (ret) {
|
|
dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
|
|
goto err_free_txport;
|
|
}
|
|
|
|
txport->non_continous_clk = vep.bus.mipi_csi2.flags &
|
|
V4L2_MBUS_CSI2_NONCONTINUOUS_CLOCK;
|
|
|
|
txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
|
|
|
|
if (vep.nr_of_link_frequencies != 1) {
|
|
ret = -EINVAL;
|
|
goto err_free_vep;
|
|
}
|
|
|
|
priv->tx_link_freq[0] = vep.link_frequencies[0];
|
|
priv->tx_data_rate = priv->tx_link_freq[0] * 2;
|
|
|
|
if (priv->tx_data_rate != MHZ(1600) &&
|
|
priv->tx_data_rate != MHZ(1200) &&
|
|
priv->tx_data_rate != MHZ(800) &&
|
|
priv->tx_data_rate != MHZ(400)) {
|
|
dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
|
|
ret = -EINVAL;
|
|
goto err_free_vep;
|
|
}
|
|
|
|
v4l2_fwnode_endpoint_free(&vep);
|
|
|
|
priv->txports[nport] = txport;
|
|
|
|
return 0;
|
|
|
|
err_free_vep:
|
|
v4l2_fwnode_endpoint_free(&vep);
|
|
err_free_txport:
|
|
kfree(txport);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 csi_tx_isr;
|
|
int ret;
|
|
|
|
ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
|
|
if (ret)
|
|
return;
|
|
|
|
if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
|
|
dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
|
|
|
|
if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
|
|
dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* RX ports
|
|
*/
|
|
|
|
static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
ret = regulator_enable(rxport->vpoc);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_disable_vpocs:
|
|
while (nport--) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
regulator_disable(rxport->vpoc);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport || !rxport->vpoc)
|
|
continue;
|
|
|
|
regulator_disable(rxport->vpoc);
|
|
}
|
|
}
|
|
|
|
static void ub960_rxport_clear_errors(struct ub960_data *priv,
|
|
unsigned int nport)
|
|
{
|
|
u8 v;
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
|
|
}
|
|
|
|
static void ub960_clear_rx_errors(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
|
|
ub960_rxport_clear_errors(priv, nport);
|
|
}
|
|
|
|
static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
|
|
unsigned int nport, s8 *strobe_pos)
|
|
{
|
|
u8 v;
|
|
u8 clk_delay, data_delay;
|
|
int ret;
|
|
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
|
|
|
|
clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
|
|
0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
|
|
|
|
data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
|
|
0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
|
|
|
|
*strobe_pos = data_delay - clk_delay;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
|
|
unsigned int nport, s8 strobe_pos)
|
|
{
|
|
u8 clk_delay, data_delay;
|
|
|
|
clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
|
|
data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
|
|
|
|
if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
|
|
clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
|
|
data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
|
|
else if (strobe_pos < 0)
|
|
clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
|
|
else if (strobe_pos > 0)
|
|
data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
|
|
UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
|
|
}
|
|
|
|
static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
|
|
s8 strobe_min, s8 strobe_max)
|
|
{
|
|
/* Convert the signed strobe pos to positive zero based value */
|
|
strobe_min -= UB960_MIN_AEQ_STROBE_POS;
|
|
strobe_max -= UB960_MIN_AEQ_STROBE_POS;
|
|
|
|
ub960_write(priv, UB960_XR_SFILTER_CFG,
|
|
((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
|
|
((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
|
|
}
|
|
|
|
static int ub960_rxport_get_eq_level(struct ub960_data *priv,
|
|
unsigned int nport, u8 *eq_level)
|
|
{
|
|
int ret;
|
|
u8 v;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
|
|
if (ret)
|
|
return ret;
|
|
|
|
*eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
|
|
(v & UB960_RR_AEQ_STATUS_STATUS_2);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_set_eq_level(struct ub960_data *priv,
|
|
unsigned int nport, u8 eq_level)
|
|
{
|
|
u8 eq_stage_1_select_value, eq_stage_2_select_value;
|
|
const unsigned int eq_stage_max = 7;
|
|
u8 v;
|
|
|
|
if (eq_level <= eq_stage_max) {
|
|
eq_stage_1_select_value = eq_level;
|
|
eq_stage_2_select_value = 0;
|
|
} else {
|
|
eq_stage_1_select_value = eq_stage_max;
|
|
eq_stage_2_select_value = eq_level - eq_stage_max;
|
|
}
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
|
|
|
|
v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
|
|
UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
|
|
v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
|
|
v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
|
|
v |= UB960_RR_AEQ_BYPASS_ENABLE;
|
|
|
|
ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
|
|
}
|
|
|
|
static void ub960_rxport_set_eq_range(struct ub960_data *priv,
|
|
unsigned int nport, u8 eq_min, u8 eq_max)
|
|
{
|
|
ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
|
|
(eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
|
|
(eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
|
|
|
|
/* Enable AEQ min setting */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
|
|
UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
|
|
UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
|
|
}
|
|
|
|
static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
/* We also set common settings here. Should be moved elsewhere. */
|
|
|
|
if (priv->strobe.manual) {
|
|
/* Disable AEQ_SFILTER_EN */
|
|
ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
|
|
UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
|
|
} else {
|
|
/* Enable SFILTER and error control */
|
|
ub960_write(priv, UB960_XR_AEQ_CTL1,
|
|
UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
|
|
UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
|
|
|
|
/* Set AEQ strobe range */
|
|
ub960_rxport_set_strobe_range(priv, priv->strobe.min,
|
|
priv->strobe.max);
|
|
}
|
|
|
|
/* The rest are port specific */
|
|
|
|
if (priv->strobe.manual)
|
|
ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
|
|
else
|
|
ub960_rxport_set_strobe_pos(priv, nport, 0);
|
|
|
|
if (rxport->eq.manual_eq) {
|
|
ub960_rxport_set_eq_level(priv, nport,
|
|
rxport->eq.manual.eq_level);
|
|
|
|
/* Enable AEQ Bypass */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
|
|
UB960_RR_AEQ_BYPASS_ENABLE,
|
|
UB960_RR_AEQ_BYPASS_ENABLE);
|
|
} else {
|
|
ub960_rxport_set_eq_range(priv, nport,
|
|
rxport->eq.aeq.eq_level_min,
|
|
rxport->eq.aeq.eq_level_max);
|
|
|
|
/* Disable AEQ Bypass */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
|
|
UB960_RR_AEQ_BYPASS_ENABLE, 0);
|
|
}
|
|
}
|
|
|
|
static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
|
|
bool *ok)
|
|
{
|
|
u8 rx_port_sts1, rx_port_sts2;
|
|
u16 parity_errors;
|
|
u8 csi_rx_sts;
|
|
u8 csi_err_cnt;
|
|
u8 bcc_sts;
|
|
int ret;
|
|
bool errors;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
|
|
&rx_port_sts1);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
|
|
*ok = false;
|
|
return 0;
|
|
}
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
|
|
&rx_port_sts2);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
|
|
&csi_err_cnt);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
|
|
&parity_errors);
|
|
if (ret)
|
|
return ret;
|
|
|
|
errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
|
|
(bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
|
|
(csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
|
|
parity_errors;
|
|
|
|
*ok = !errors;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Wait for the RX ports to lock, have no errors and have stable strobe position
|
|
* and EQ level.
|
|
*/
|
|
static int ub960_rxport_wait_locks(struct ub960_data *priv,
|
|
unsigned long port_mask,
|
|
unsigned int *lock_mask)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned long timeout;
|
|
unsigned int link_ok_mask;
|
|
unsigned int missing;
|
|
unsigned int loops;
|
|
u8 nport;
|
|
int ret;
|
|
|
|
if (port_mask == 0) {
|
|
if (lock_mask)
|
|
*lock_mask = 0;
|
|
return 0;
|
|
}
|
|
|
|
if (port_mask >= BIT(priv->hw_data->num_rxports))
|
|
return -EINVAL;
|
|
|
|
timeout = jiffies + msecs_to_jiffies(1000);
|
|
loops = 0;
|
|
link_ok_mask = 0;
|
|
|
|
while (time_before(jiffies, timeout)) {
|
|
missing = 0;
|
|
|
|
for_each_set_bit(nport, &port_mask,
|
|
priv->hw_data->num_rxports) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
bool ok;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = ub960_rxport_link_ok(priv, nport, &ok);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* We want the link to be ok for two consecutive loops,
|
|
* as a link could get established just before our test
|
|
* and drop soon after.
|
|
*/
|
|
if (!ok || !(link_ok_mask & BIT(nport)))
|
|
missing++;
|
|
|
|
if (ok)
|
|
link_ok_mask |= BIT(nport);
|
|
else
|
|
link_ok_mask &= ~BIT(nport);
|
|
}
|
|
|
|
loops++;
|
|
|
|
if (missing == 0)
|
|
break;
|
|
|
|
msleep(50);
|
|
}
|
|
|
|
if (lock_mask)
|
|
*lock_mask = link_ok_mask;
|
|
|
|
dev_dbg(dev, "Wait locks done in %u loops\n", loops);
|
|
for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
s8 strobe_pos, eq_level;
|
|
u16 v;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
if (!(link_ok_mask & BIT(nport))) {
|
|
dev_dbg(dev, "\trx%u: not locked\n", nport);
|
|
continue;
|
|
}
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
|
|
|
|
ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
|
|
if (ret)
|
|
return ret;
|
|
|
|
dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
|
|
nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int mult;
|
|
unsigned int div;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
mult = 1;
|
|
div = 10;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
mult = 2;
|
|
div = 1;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
mult = 2;
|
|
div = 5;
|
|
break;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
|
|
return clk_get_rate(priv->refclk) * mult / div;
|
|
}
|
|
|
|
static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
return 2359400;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
return 47187500;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
return 9437500;
|
|
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
struct device *dev = &priv->client->dev;
|
|
struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
|
|
struct i2c_board_info ser_info = {
|
|
.of_node = to_of_node(rxport->ser.fwnode),
|
|
.fwnode = rxport->ser.fwnode,
|
|
.platform_data = ser_pdata,
|
|
};
|
|
|
|
ser_pdata->port = nport;
|
|
ser_pdata->atr = priv->atr;
|
|
if (priv->hw_data->is_ub9702)
|
|
ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
|
|
else
|
|
ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
|
|
|
|
/*
|
|
* The serializer is added under the same i2c adapter as the
|
|
* deserializer. This is not quite right, as the serializer is behind
|
|
* the FPD-Link.
|
|
*/
|
|
ser_info.addr = rxport->ser.alias;
|
|
rxport->ser.client =
|
|
i2c_new_client_device(priv->client->adapter, &ser_info);
|
|
if (IS_ERR(rxport->ser.client)) {
|
|
dev_err(dev, "rx%u: cannot add %s i2c device", nport,
|
|
ser_info.type);
|
|
return PTR_ERR(rxport->ser.client);
|
|
}
|
|
|
|
dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
|
|
nport, rxport->ser.client->addr,
|
|
rxport->ser.client->adapter->nr, rxport->ser.client->addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
i2c_unregister_device(rxport->ser.client);
|
|
rxport->ser.client = NULL;
|
|
}
|
|
|
|
/* Add serializer i2c devices for all initialized ports */
|
|
static int ub960_rxport_add_serializers(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = ub960_rxport_add_serializer(priv, nport);
|
|
if (ret)
|
|
goto err_remove_sers;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_remove_sers:
|
|
while (nport--) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ub960_rxport_remove_serializer(priv, nport);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_rxport_remove_serializers(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ub960_rxport_remove_serializer(priv, nport);
|
|
}
|
|
}
|
|
|
|
static void ub960_init_tx_port(struct ub960_data *priv,
|
|
struct ub960_txport *txport)
|
|
{
|
|
unsigned int nport = txport->nport;
|
|
u8 csi_ctl = 0;
|
|
|
|
/*
|
|
* From the datasheet: "initial CSI Skew-Calibration
|
|
* sequence [...] should be set when operating at 1.6 Gbps"
|
|
*/
|
|
if (priv->tx_data_rate == MHZ(1600))
|
|
csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
|
|
|
|
csi_ctl |= (4 - txport->num_data_lanes) << 4;
|
|
|
|
if (!txport->non_continous_clk)
|
|
csi_ctl |= UB960_TR_CSI_CTL_CSI_CONTS_CLOCK;
|
|
|
|
ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
|
|
}
|
|
|
|
static int ub960_init_tx_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
u8 speed_select;
|
|
u8 pll_div;
|
|
|
|
/* TX ports */
|
|
|
|
switch (priv->tx_data_rate) {
|
|
case MHZ(1600):
|
|
default:
|
|
speed_select = 0;
|
|
pll_div = 0x10;
|
|
break;
|
|
case MHZ(1200):
|
|
speed_select = 1;
|
|
break;
|
|
case MHZ(800):
|
|
speed_select = 2;
|
|
pll_div = 0x10;
|
|
break;
|
|
case MHZ(400):
|
|
speed_select = 3;
|
|
pll_div = 0x10;
|
|
break;
|
|
}
|
|
|
|
ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
|
|
|
|
if (priv->hw_data->is_ub9702) {
|
|
ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
|
|
|
|
switch (priv->tx_data_rate) {
|
|
case MHZ(1600):
|
|
default:
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
|
|
break;
|
|
case MHZ(800):
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
|
|
break;
|
|
case MHZ(400):
|
|
ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
if (!txport)
|
|
continue;
|
|
|
|
ub960_init_tx_port(priv, txport);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub960(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u32 bc_freq_val;
|
|
|
|
/*
|
|
* Back channel frequency select.
|
|
* Override FREQ_SELECT from the strap.
|
|
* 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
|
|
* 2 - 10 Mbps
|
|
* 6 - 50 Mbps (DS90UB953-Q1)
|
|
*
|
|
* Note that changing this setting will result in some errors on the back
|
|
* channel for a short period of time.
|
|
*/
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
|
|
bc_freq_val);
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
/* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
|
|
UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
|
|
0x3);
|
|
|
|
/*
|
|
* RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
|
|
*/
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
|
|
UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
|
|
0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
return;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
/* CSI-2 Mode (DS90UB953-Q1 compatible) */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
|
|
0x0);
|
|
|
|
break;
|
|
}
|
|
|
|
/* LV_POLARITY & FV_POLARITY */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
|
|
rxport->lv_fv_pol);
|
|
|
|
/* Enable all interrupt sources from this port */
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
|
|
|
|
/* Enable I2C_PASS_THROUGH */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
|
|
|
|
/* Enable I2C communication to the serializer via the alias addr */
|
|
ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
|
|
rxport->ser.alias << 1);
|
|
|
|
/* Configure EQ related settings */
|
|
ub960_rxport_config_eq(priv, nport);
|
|
|
|
/* Enable RX port */
|
|
ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u8 bc_freq_val;
|
|
u8 fpd_func_mode;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 5;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 4;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
fpd_func_mode = 6;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
fpd_func_mode = 2;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
fpd_func_mode = 2;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
|
|
bc_freq_val);
|
|
ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
|
|
|
|
/* set serdes_eq_mode = 1 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
|
|
|
|
/* enable serdes driver */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
|
|
|
|
/* set serdes_eq_offset=4 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
|
|
|
|
/* init default serdes_eq_max in 0xa9 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
|
|
|
|
/* init serdes_eq_min in 0xaa */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
|
|
|
|
/* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
|
|
ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
|
|
BIT(3), BIT(3));
|
|
|
|
/* RX port to half-rate */
|
|
ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
|
|
BIT(nport * 2));
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
bool first_time_power_up = true;
|
|
|
|
if (first_time_power_up) {
|
|
u8 v;
|
|
|
|
/* AEQ init */
|
|
ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
|
|
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
|
|
}
|
|
|
|
/* enable serdes_eq_ctl2 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
|
|
|
|
/* enable serdes_eq_ctl1 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
|
|
|
|
/* enable serdes_eq_en */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
|
|
|
|
/* disable serdes_eq_override */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
|
|
|
|
/* disable serdes_gain_override */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
u8 bc_freq_val;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_LF:
|
|
bc_freq_val = 0;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
bc_freq_val = 6;
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
bc_freq_val = 2;
|
|
break;
|
|
|
|
default:
|
|
return;
|
|
}
|
|
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
|
|
bc_freq_val);
|
|
|
|
/* FPD4 Sync Mode */
|
|
ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
|
|
|
|
/* add serdes_eq_offset of 4 */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
|
|
|
|
/* FPD4 serdes_start_eq in 0x27: assign default */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
|
|
/* FPD4 serdes_end_eq in 0x28: assign default */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
|
|
|
|
/* set serdes_driver_mode into FPD IV mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
|
|
/* set FPD PBC drv into FPD IV mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
|
|
|
|
/* set serdes_system_init to 0x2f */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
|
|
/* set serdes_system_rst in reset mode */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
|
|
|
|
/* RX port to 7.55G mode */
|
|
ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
|
|
0 << (nport * 2));
|
|
|
|
ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
|
|
}
|
|
|
|
static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
unsigned int nport = rxport->nport;
|
|
|
|
if (rxport->cdr_mode == RXPORT_CDR_FPD3)
|
|
ub960_init_rx_port_ub9702_fpd3(priv, rxport);
|
|
else /* RXPORT_CDR_FPD4 */
|
|
ub960_init_rx_port_ub9702_fpd4(priv, rxport);
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
/*
|
|
* RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
|
|
* 0b10 : 8-bit processing using upper 8 bits
|
|
*/
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
|
|
0x3 << 6, 0x2 << 6);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
return;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
|
|
break;
|
|
}
|
|
|
|
/* LV_POLARITY & FV_POLARITY */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
|
|
rxport->lv_fv_pol);
|
|
|
|
/* Enable all interrupt sources from this port */
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
|
|
ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
|
|
|
|
/* Enable I2C_PASS_THROUGH */
|
|
ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
|
|
UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
|
|
|
|
/* Enable I2C communication to the serializer via the alias addr */
|
|
ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
|
|
rxport->ser.alias << 1);
|
|
|
|
/* Enable RX port */
|
|
ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
|
|
|
|
if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
|
|
/* unreset 960 AEQ */
|
|
ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
|
|
}
|
|
}
|
|
|
|
static int ub960_init_rx_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
if (priv->hw_data->is_ub9702)
|
|
ub960_init_rx_port_ub9702(priv, rxport);
|
|
else
|
|
ub960_init_rx_port_ub960(priv, rxport);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 rx_port_sts1;
|
|
u8 rx_port_sts2;
|
|
u8 csi_rx_sts;
|
|
u8 bcc_sts;
|
|
int ret = 0;
|
|
|
|
/* Read interrupts (also clears most of them) */
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
|
|
&rx_port_sts1);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
|
|
&rx_port_sts2);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
|
|
&csi_rx_sts);
|
|
if (!ret)
|
|
ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
|
|
&bcc_sts);
|
|
|
|
if (ret)
|
|
return;
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
|
|
&v);
|
|
if (!ret)
|
|
dev_err(dev, "rx%u parity errors: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
|
|
dev_err(dev, "rx%u BCC CRC error\n", nport);
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
|
|
dev_err(dev, "rx%u BCC SEQ error\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
|
|
dev_err(dev, "rx%u line length unstable\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
|
|
dev_err(dev, "rx%u FPD3 encode error\n", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
|
|
dev_err(dev, "rx%u buffer error\n", nport);
|
|
|
|
if (csi_rx_sts)
|
|
dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
|
|
dev_err(dev, "rx%u CSI ECC1 error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
|
|
dev_err(dev, "rx%u CSI ECC2 error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
|
|
dev_err(dev, "rx%u CSI checksum error\n", nport);
|
|
|
|
if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
|
|
dev_err(dev, "rx%u CSI length error\n", nport);
|
|
|
|
if (bcc_sts)
|
|
dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
|
|
dev_err(dev, "rx%u BCC response error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
|
|
dev_err(dev, "rx%u BCC slave timeout", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
|
|
dev_err(dev, "rx%u BCC slave error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
|
|
dev_err(dev, "rx%u BCC master timeout", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
|
|
dev_err(dev, "rx%u BCC master error", nport);
|
|
|
|
if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
|
|
dev_err(dev, "rx%u BCC sequence error", nport);
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
|
|
if (!ret)
|
|
dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
|
|
u16 v;
|
|
|
|
ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
|
|
&v);
|
|
if (!ret)
|
|
dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
|
|
}
|
|
|
|
if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
|
|
dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
|
|
(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
|
|
"locked" :
|
|
"unlocked",
|
|
(rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
|
|
"passed" :
|
|
"not passed",
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
|
|
"no clock" :
|
|
"clock ok",
|
|
(rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
|
|
"stable freq" :
|
|
"unstable freq");
|
|
}
|
|
}
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* V4L2
|
|
*/
|
|
|
|
/*
|
|
* The current implementation only supports a simple VC mapping, where all VCs
|
|
* from a one RX port will be mapped to the same VC. Also, the hardware
|
|
* dictates that all streams from an RX port must go to a single TX port.
|
|
*
|
|
* This function decides the target VC numbers for each RX port with a simple
|
|
* algorithm, so that for each TX port, we get VC numbers starting from 0,
|
|
* and counting up.
|
|
*
|
|
* E.g. if all four RX ports are in use, of which the first two go to the
|
|
* first TX port and the secont two go to the second TX port, we would get
|
|
* the following VCs for the four RX ports: 0, 1, 0, 1.
|
|
*
|
|
* TODO: implement a more sophisticated VC mapping. As the driver cannot know
|
|
* what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
|
|
* probably needs to be somehow configurable. Device tree?
|
|
*/
|
|
static void ub960_get_vc_maps(struct ub960_data *priv,
|
|
struct v4l2_subdev_state *state, u8 *vc)
|
|
{
|
|
u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
u8 handled_mask = 0;
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
unsigned int rx, tx;
|
|
|
|
rx = ub960_pad_to_port(priv, route->sink_pad);
|
|
if (BIT(rx) & handled_mask)
|
|
continue;
|
|
|
|
tx = ub960_pad_to_port(priv, route->source_pad);
|
|
|
|
vc[rx] = cur_vc[tx]++;
|
|
handled_mask |= BIT(rx);
|
|
}
|
|
}
|
|
|
|
static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "enable TX port %u\n", nport);
|
|
|
|
return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE);
|
|
}
|
|
|
|
static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "disable TX port %u\n", nport);
|
|
|
|
ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
|
|
UB960_TR_CSI_CTL_CSI_ENABLE, 0);
|
|
}
|
|
|
|
static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "enable RX port %u\n", nport);
|
|
|
|
/* Enable forwarding */
|
|
return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
|
|
}
|
|
|
|
static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
dev_dbg(dev, "disable RX port %u\n", nport);
|
|
|
|
/* Disable forwarding */
|
|
ub960_update_bits(priv, UB960_SR_FWD_CTL1,
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport),
|
|
UB960_SR_FWD_CTL1_PORT_DIS(nport));
|
|
}
|
|
|
|
/*
|
|
* The driver only supports using a single VC for each source. This function
|
|
* checks that each source only provides streams using a single VC.
|
|
*/
|
|
static int ub960_validate_stream_vcs(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
struct v4l2_mbus_frame_desc desc;
|
|
int ret;
|
|
u8 vc;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
|
|
rxport->source.pad, &desc);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
|
|
continue;
|
|
|
|
if (desc.num_entries == 0)
|
|
continue;
|
|
|
|
vc = desc.entry[0].bus.csi2.vc;
|
|
|
|
for (i = 1; i < desc.num_entries; i++) {
|
|
if (vc == desc.entry[i].bus.csi2.vc)
|
|
continue;
|
|
|
|
dev_err(&priv->client->dev,
|
|
"rx%u: source with multiple virtual-channels is not supported\n",
|
|
nport);
|
|
return -ENODEV;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
|
|
struct v4l2_subdev_state *state)
|
|
{
|
|
u8 fwd_ctl;
|
|
struct {
|
|
u32 num_streams;
|
|
u8 pixel_dt;
|
|
u8 meta_dt;
|
|
u32 meta_lines;
|
|
u32 tx_port;
|
|
} rx_data[UB960_MAX_RX_NPORTS] = {};
|
|
u8 vc_map[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
ret = ub960_validate_stream_vcs(priv);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ub960_get_vc_maps(priv, state, vc_map);
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
struct ub960_rxport *rxport;
|
|
struct ub960_txport *txport;
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
const struct ub960_format_info *ub960_fmt;
|
|
unsigned int nport;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
rxport = priv->rxports[nport];
|
|
if (!rxport)
|
|
return -EINVAL;
|
|
|
|
txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
|
|
if (!txport)
|
|
return -EINVAL;
|
|
|
|
rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
|
|
|
|
rx_data[nport].num_streams++;
|
|
|
|
/* For the rest, we are only interested in parallel busses */
|
|
if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
|
|
rxport->rx_mode == RXPORT_MODE_CSI2_ASYNC)
|
|
continue;
|
|
|
|
if (rx_data[nport].num_streams > 2)
|
|
return -EPIPE;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state,
|
|
route->sink_pad,
|
|
route->sink_stream);
|
|
if (!fmt)
|
|
return -EPIPE;
|
|
|
|
ub960_fmt = ub960_find_format(fmt->code);
|
|
if (!ub960_fmt)
|
|
return -EPIPE;
|
|
|
|
if (ub960_fmt->meta) {
|
|
if (fmt->height > 3) {
|
|
dev_err(&priv->client->dev,
|
|
"rx%u: unsupported metadata height %u\n",
|
|
nport, fmt->height);
|
|
return -EPIPE;
|
|
}
|
|
|
|
rx_data[nport].meta_dt = ub960_fmt->datatype;
|
|
rx_data[nport].meta_lines = fmt->height;
|
|
} else {
|
|
rx_data[nport].pixel_dt = ub960_fmt->datatype;
|
|
}
|
|
}
|
|
|
|
/* Configure RX ports */
|
|
|
|
/*
|
|
* Keep all port forwardings disabled by default. Forwarding will be
|
|
* enabled in ub960_enable_rx_port.
|
|
*/
|
|
fwd_ctl = GENMASK(7, 4);
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
u8 vc = vc_map[nport];
|
|
|
|
if (rx_data[nport].num_streams == 0)
|
|
continue;
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
|
|
rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
|
|
|
|
ub960_rxport_write(priv, rxport->nport,
|
|
UB960_RR_RAW_EMBED_DTYPE,
|
|
(rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
|
|
rx_data[nport].meta_dt);
|
|
|
|
break;
|
|
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
/* Not implemented */
|
|
break;
|
|
|
|
case RXPORT_MODE_CSI2_SYNC:
|
|
case RXPORT_MODE_CSI2_ASYNC:
|
|
if (!priv->hw_data->is_ub9702) {
|
|
/* Map all VCs from this port to the same VC */
|
|
ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
|
|
(vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
|
|
} else {
|
|
unsigned int i;
|
|
|
|
/* Map all VCs from this port to VC(nport) */
|
|
for (i = 0; i < 8; i++)
|
|
ub960_rxport_write(priv, nport,
|
|
UB960_RR_VC_ID_MAP(i),
|
|
nport);
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
if (rx_data[nport].tx_port == 1)
|
|
fwd_ctl |= BIT(nport); /* forward to TX1 */
|
|
else
|
|
fwd_ctl &= ~BIT(nport); /* forward to TX0 */
|
|
}
|
|
|
|
ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_update_streaming_status(struct ub960_data *priv)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < UB960_MAX_NPORTS; i++) {
|
|
if (priv->stream_enable_mask[i])
|
|
break;
|
|
}
|
|
|
|
priv->streaming = i < UB960_MAX_NPORTS;
|
|
}
|
|
|
|
static int ub960_enable_streams(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state, u32 source_pad,
|
|
u64 source_streams_mask)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int failed_port;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
if (!priv->streaming) {
|
|
dev_dbg(dev, "Prepare for streaming\n");
|
|
ret = ub960_configure_ports_for_streaming(priv, state);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/* Enable TX port if not yet enabled */
|
|
if (!priv->stream_enable_mask[source_pad]) {
|
|
ret = ub960_enable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
priv->stream_enable_mask[source_pad] |= source_streams_mask;
|
|
|
|
/* Collect sink streams per pad which we need to enable */
|
|
for_each_active_route(&state->routing, route) {
|
|
if (route->source_pad != source_pad)
|
|
continue;
|
|
|
|
if (!(source_streams_mask & BIT_ULL(route->source_stream)))
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
sink_streams[nport] |= BIT_ULL(route->sink_stream);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
/* Enable the RX port if not yet enabled */
|
|
if (!priv->stream_enable_mask[nport]) {
|
|
ret = ub960_enable_rx_port(priv, nport);
|
|
if (ret) {
|
|
failed_port = nport;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
priv->stream_enable_mask[nport] |= sink_streams[nport];
|
|
|
|
dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_enable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret) {
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
|
|
failed_port = nport;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
priv->streaming = true;
|
|
|
|
return 0;
|
|
|
|
err:
|
|
for (nport = 0; nport < failed_port; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_disable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret)
|
|
dev_err(dev, "Failed to disable streams: %d\n", ret);
|
|
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
/* Disable RX port if no active streams */
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
}
|
|
|
|
priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
|
|
|
|
if (!priv->stream_enable_mask[source_pad])
|
|
ub960_disable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
|
|
ub960_update_streaming_status(priv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_disable_streams(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
u32 source_pad, u64 source_streams_mask)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
|
|
struct v4l2_subdev_route *route;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
/* Collect sink streams per pad which we need to disable */
|
|
for_each_active_route(&state->routing, route) {
|
|
if (route->source_pad != source_pad)
|
|
continue;
|
|
|
|
if (!(source_streams_mask & BIT_ULL(route->source_stream)))
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
sink_streams[nport] |= BIT_ULL(route->sink_stream);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
if (!sink_streams[nport])
|
|
continue;
|
|
|
|
dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
|
|
sink_streams[nport]);
|
|
|
|
ret = v4l2_subdev_disable_streams(
|
|
priv->rxports[nport]->source.sd,
|
|
priv->rxports[nport]->source.pad,
|
|
sink_streams[nport]);
|
|
if (ret)
|
|
dev_err(dev, "Failed to disable streams: %d\n", ret);
|
|
|
|
priv->stream_enable_mask[nport] &= ~sink_streams[nport];
|
|
|
|
/* Disable RX port if no active streams */
|
|
if (!priv->stream_enable_mask[nport])
|
|
ub960_disable_rx_port(priv, nport);
|
|
}
|
|
|
|
/* Disable TX port if no active streams */
|
|
|
|
priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
|
|
|
|
if (!priv->stream_enable_mask[source_pad])
|
|
ub960_disable_tx_port(priv,
|
|
ub960_pad_to_port(priv, source_pad));
|
|
|
|
ub960_update_streaming_status(priv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int _ub960_set_routing(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
struct v4l2_subdev_krouting *routing)
|
|
{
|
|
static const struct v4l2_mbus_framefmt format = {
|
|
.width = 640,
|
|
.height = 480,
|
|
.code = MEDIA_BUS_FMT_UYVY8_1X16,
|
|
.field = V4L2_FIELD_NONE,
|
|
.colorspace = V4L2_COLORSPACE_SRGB,
|
|
.ycbcr_enc = V4L2_YCBCR_ENC_601,
|
|
.quantization = V4L2_QUANTIZATION_LIM_RANGE,
|
|
.xfer_func = V4L2_XFER_FUNC_SRGB,
|
|
};
|
|
int ret;
|
|
|
|
/*
|
|
* Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
|
|
* frame desc is made dynamically allocated.
|
|
*/
|
|
|
|
if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
|
|
return -E2BIG;
|
|
|
|
ret = v4l2_subdev_routing_validate(sd, routing,
|
|
V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
|
|
V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_set_routing(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
enum v4l2_subdev_format_whence which,
|
|
struct v4l2_subdev_krouting *routing)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
|
|
return -EBUSY;
|
|
|
|
return _ub960_set_routing(sd, state, routing);
|
|
}
|
|
|
|
static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
|
|
struct v4l2_mbus_frame_desc *fd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct v4l2_subdev_route *route;
|
|
struct v4l2_subdev_state *state;
|
|
int ret = 0;
|
|
struct device *dev = &priv->client->dev;
|
|
u8 vc_map[UB960_MAX_RX_NPORTS] = {};
|
|
|
|
if (!ub960_pad_is_source(priv, pad))
|
|
return -EINVAL;
|
|
|
|
memset(fd, 0, sizeof(*fd));
|
|
|
|
fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
|
|
|
|
state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
|
|
|
|
ub960_get_vc_maps(priv, state, vc_map);
|
|
|
|
for_each_active_route(&state->routing, route) {
|
|
struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
|
|
struct v4l2_mbus_frame_desc source_fd;
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
|
|
if (route->source_pad != pad)
|
|
continue;
|
|
|
|
nport = ub960_pad_to_port(priv, route->sink_pad);
|
|
|
|
ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
|
|
get_frame_desc,
|
|
priv->rxports[nport]->source.pad,
|
|
&source_fd);
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Failed to get source frame desc for pad %u\n",
|
|
route->sink_pad);
|
|
goto out_unlock;
|
|
}
|
|
|
|
for (i = 0; i < source_fd.num_entries; i++) {
|
|
if (source_fd.entry[i].stream == route->sink_stream) {
|
|
source_entry = &source_fd.entry[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!source_entry) {
|
|
dev_err(dev,
|
|
"Failed to find stream from source frame desc\n");
|
|
ret = -EPIPE;
|
|
goto out_unlock;
|
|
}
|
|
|
|
fd->entry[fd->num_entries].stream = route->source_stream;
|
|
fd->entry[fd->num_entries].flags = source_entry->flags;
|
|
fd->entry[fd->num_entries].length = source_entry->length;
|
|
fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
|
|
|
|
fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
|
|
|
|
if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
|
|
fd->entry[fd->num_entries].bus.csi2.dt =
|
|
source_entry->bus.csi2.dt;
|
|
} else {
|
|
const struct ub960_format_info *ub960_fmt;
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state, pad,
|
|
route->source_stream);
|
|
|
|
if (!fmt) {
|
|
ret = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
ub960_fmt = ub960_find_format(fmt->code);
|
|
if (!ub960_fmt) {
|
|
dev_err(dev, "Unable to find format\n");
|
|
ret = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
fd->entry[fd->num_entries].bus.csi2.dt =
|
|
ub960_fmt->datatype;
|
|
}
|
|
|
|
fd->num_entries++;
|
|
}
|
|
|
|
out_unlock:
|
|
v4l2_subdev_unlock_state(state);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_set_fmt(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state,
|
|
struct v4l2_subdev_format *format)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct v4l2_mbus_framefmt *fmt;
|
|
|
|
if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
|
|
return -EBUSY;
|
|
|
|
/* No transcoding, source and sink formats must match. */
|
|
if (ub960_pad_is_source(priv, format->pad))
|
|
return v4l2_subdev_get_fmt(sd, state, format);
|
|
|
|
/*
|
|
* Default to the first format if the requested media bus code isn't
|
|
* supported.
|
|
*/
|
|
if (!ub960_find_format(format->format.code))
|
|
format->format.code = ub960_formats[0].code;
|
|
|
|
fmt = v4l2_subdev_state_get_stream_format(state, format->pad,
|
|
format->stream);
|
|
if (!fmt)
|
|
return -EINVAL;
|
|
|
|
*fmt = format->format;
|
|
|
|
fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
|
|
format->stream);
|
|
if (!fmt)
|
|
return -EINVAL;
|
|
|
|
*fmt = format->format;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_init_cfg(struct v4l2_subdev *sd,
|
|
struct v4l2_subdev_state *state)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
struct v4l2_subdev_route routes[] = {
|
|
{
|
|
.sink_pad = 0,
|
|
.sink_stream = 0,
|
|
.source_pad = priv->hw_data->num_rxports,
|
|
.source_stream = 0,
|
|
.flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
|
|
},
|
|
};
|
|
|
|
struct v4l2_subdev_krouting routing = {
|
|
.num_routes = ARRAY_SIZE(routes),
|
|
.routes = routes,
|
|
};
|
|
|
|
return _ub960_set_routing(sd, state, &routing);
|
|
}
|
|
|
|
static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
|
|
.enable_streams = ub960_enable_streams,
|
|
.disable_streams = ub960_disable_streams,
|
|
|
|
.set_routing = ub960_set_routing,
|
|
.get_frame_desc = ub960_get_frame_desc,
|
|
|
|
.get_fmt = v4l2_subdev_get_fmt,
|
|
.set_fmt = ub960_set_fmt,
|
|
|
|
.init_cfg = ub960_init_cfg,
|
|
};
|
|
|
|
static int ub960_log_status(struct v4l2_subdev *sd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_subdev_state *state;
|
|
unsigned int nport;
|
|
unsigned int i;
|
|
u16 v16 = 0;
|
|
u8 v = 0;
|
|
u8 id[UB960_SR_FPD3_RX_ID_LEN];
|
|
|
|
state = v4l2_subdev_lock_and_get_active_state(sd);
|
|
|
|
for (i = 0; i < sizeof(id); i++)
|
|
ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
|
|
|
|
dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
dev_info(dev, "TX %u\n", nport);
|
|
|
|
if (!txport) {
|
|
dev_info(dev, "\tNot initialized\n");
|
|
continue;
|
|
}
|
|
|
|
ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
|
|
dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
|
|
v & (u8)BIT(0));
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tframe counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tframe error counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tline counter %u\n", v16);
|
|
|
|
ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
|
|
dev_info(dev, "\tline error counter %u\n", v16);
|
|
}
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
u8 eq_level;
|
|
s8 strobe_pos;
|
|
unsigned int i;
|
|
|
|
dev_info(dev, "RX %u\n", nport);
|
|
|
|
if (!rxport) {
|
|
dev_info(dev, "\tNot initialized\n");
|
|
continue;
|
|
}
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
|
|
|
|
if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
|
|
dev_info(dev, "\tLocked\n");
|
|
else
|
|
dev_info(dev, "\tNot locked\n");
|
|
|
|
dev_info(dev, "\trx_port_sts1 %#02x\n", v);
|
|
ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
|
|
dev_info(dev, "\trx_port_sts2 %#02x\n", v);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
|
|
dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
|
|
dev_info(dev, "\tparity errors %u\n", v16);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
|
|
dev_info(dev, "\tlines per frame %u\n", v16);
|
|
|
|
ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
|
|
dev_info(dev, "\tbytes per line %u\n", v16);
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
|
|
dev_info(dev, "\tcsi_err_counter %u\n", v);
|
|
|
|
/* Strobe */
|
|
|
|
ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
|
|
|
|
dev_info(dev, "\t%s strobe\n",
|
|
(v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
|
|
"Manual");
|
|
|
|
if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
|
|
ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
|
|
|
|
dev_info(dev, "\tStrobe range [%d, %d]\n",
|
|
((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
|
|
((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
|
|
}
|
|
|
|
ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
|
|
|
|
dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
|
|
|
|
/* EQ */
|
|
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
|
|
|
|
dev_info(dev, "\t%s EQ\n",
|
|
(v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
|
|
"Adaptive");
|
|
|
|
if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
|
|
ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
|
|
|
|
dev_info(dev, "\tEQ range [%u, %u]\n",
|
|
(v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
|
|
(v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
|
|
}
|
|
|
|
if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
|
|
dev_info(dev, "\tEQ level %u\n", eq_level);
|
|
|
|
/* GPIOs */
|
|
for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
|
|
u8 ctl_reg;
|
|
u8 ctl_shift;
|
|
|
|
ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
|
|
ctl_shift = (i % 2) * 4;
|
|
|
|
ub960_rxport_read(priv, nport, ctl_reg, &v);
|
|
|
|
dev_info(dev, "\tGPIO%u: mode %u\n", i,
|
|
(v >> ctl_shift) & 0xf);
|
|
}
|
|
}
|
|
|
|
v4l2_subdev_unlock_state(state);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
|
|
.log_status = ub960_log_status,
|
|
.subscribe_event = v4l2_ctrl_subdev_subscribe_event,
|
|
.unsubscribe_event = v4l2_event_subdev_unsubscribe,
|
|
};
|
|
|
|
static const struct v4l2_subdev_ops ub960_subdev_ops = {
|
|
.core = &ub960_subdev_core_ops,
|
|
.pad = &ub960_pad_ops,
|
|
};
|
|
|
|
static const struct media_entity_operations ub960_entity_ops = {
|
|
.get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
|
|
.link_validate = v4l2_subdev_link_validate,
|
|
.has_pad_interdep = v4l2_subdev_has_pad_interdep,
|
|
};
|
|
|
|
/* -----------------------------------------------------------------------------
|
|
* Core
|
|
*/
|
|
|
|
static irqreturn_t ub960_handle_events(int irq, void *arg)
|
|
{
|
|
struct ub960_data *priv = arg;
|
|
unsigned int i;
|
|
u8 int_sts;
|
|
u8 fwd_sts;
|
|
int ret;
|
|
|
|
ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
|
|
if (ret || !int_sts)
|
|
return IRQ_NONE;
|
|
|
|
dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
|
|
|
|
ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
|
|
if (ret)
|
|
return IRQ_NONE;
|
|
|
|
dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
|
|
|
|
for (i = 0; i < priv->hw_data->num_txports; i++) {
|
|
if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
|
|
ub960_csi_handle_events(priv, i);
|
|
}
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
if (!priv->rxports[i])
|
|
continue;
|
|
|
|
if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
|
|
ub960_rxport_handle_events(priv, i);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void ub960_handler_work(struct work_struct *work)
|
|
{
|
|
struct delayed_work *dwork = to_delayed_work(work);
|
|
struct ub960_data *priv =
|
|
container_of(dwork, struct ub960_data, poll_work);
|
|
|
|
ub960_handle_events(0, priv);
|
|
|
|
schedule_delayed_work(&priv->poll_work,
|
|
msecs_to_jiffies(UB960_POLL_TIME_MS));
|
|
}
|
|
|
|
static void ub960_txport_free_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
struct ub960_txport *txport = priv->txports[nport];
|
|
|
|
if (!txport)
|
|
continue;
|
|
|
|
kfree(txport);
|
|
priv->txports[nport] = NULL;
|
|
}
|
|
}
|
|
|
|
static void ub960_rxport_free_ports(struct ub960_data *priv)
|
|
{
|
|
unsigned int nport;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
fwnode_handle_put(rxport->source.ep_fwnode);
|
|
fwnode_handle_put(rxport->ser.fwnode);
|
|
|
|
kfree(rxport);
|
|
priv->rxports[nport] = NULL;
|
|
}
|
|
}
|
|
|
|
static int
|
|
ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
|
|
struct fwnode_handle *link_fwnode,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int nport = rxport->nport;
|
|
u32 rx_mode;
|
|
u32 cdr_mode;
|
|
s32 strobe_pos;
|
|
u32 eq_level;
|
|
u32 ser_i2c_alias;
|
|
int ret;
|
|
|
|
cdr_mode = RXPORT_CDR_FPD3;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
|
|
if (ret < 0 && ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,cdr-mode", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (cdr_mode > RXPORT_CDR_LAST) {
|
|
dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
|
|
dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
|
|
return -EINVAL;
|
|
}
|
|
|
|
rxport->cdr_mode = cdr_mode;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
|
|
if (ret < 0) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,rx-mode", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (rx_mode > RXPORT_MODE_LAST) {
|
|
dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
|
|
return -EINVAL;
|
|
}
|
|
|
|
switch (rx_mode) {
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
|
|
rx_mode);
|
|
return -EINVAL;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
rxport->rx_mode = rx_mode;
|
|
|
|
/* EQ & Strobe related */
|
|
|
|
/* Defaults */
|
|
rxport->eq.manual_eq = false;
|
|
rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
|
|
rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
|
|
&strobe_pos);
|
|
if (ret) {
|
|
if (ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,strobe-pos", ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
|
|
strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
|
|
dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
|
|
nport, strobe_pos);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* NOTE: ignored unless global manual strobe pos is also set */
|
|
rxport->eq.strobe_pos = strobe_pos;
|
|
if (!priv->strobe.manual)
|
|
dev_warn(dev,
|
|
"rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
|
|
nport);
|
|
}
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
|
|
if (ret) {
|
|
if (ret != -EINVAL) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"ti,eq-level", ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
if (eq_level > UB960_MAX_EQ_LEVEL) {
|
|
dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
|
|
nport, eq_level);
|
|
return -EINVAL;
|
|
}
|
|
|
|
rxport->eq.manual_eq = true;
|
|
rxport->eq.manual.eq_level = eq_level;
|
|
}
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
|
|
&ser_i2c_alias);
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
|
|
"i2c-alias", ret);
|
|
return ret;
|
|
}
|
|
rxport->ser.alias = ser_i2c_alias;
|
|
|
|
rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
|
|
if (!rxport->ser.fwnode) {
|
|
dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
|
|
struct fwnode_handle *ep_fwnode,
|
|
struct ub960_rxport *rxport)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct v4l2_fwnode_endpoint vep = {};
|
|
unsigned int nport = rxport->nport;
|
|
bool hsync_hi;
|
|
bool vsync_hi;
|
|
int ret;
|
|
|
|
rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
|
|
if (!rxport->source.ep_fwnode) {
|
|
dev_err(dev, "rx%u: no remote endpoint\n", nport);
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* We currently have properties only for RAW modes */
|
|
|
|
switch (rxport->rx_mode) {
|
|
case RXPORT_MODE_RAW10:
|
|
case RXPORT_MODE_RAW12_HF:
|
|
case RXPORT_MODE_RAW12_LF:
|
|
break;
|
|
default:
|
|
return 0;
|
|
}
|
|
|
|
vep.bus_type = V4L2_MBUS_PARALLEL;
|
|
ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
|
|
goto err_put_source_ep_fwnode;
|
|
}
|
|
|
|
hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
|
|
vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
|
|
|
|
/* LineValid and FrameValid are inverse to the h/vsync active */
|
|
rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
|
|
(vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
|
|
|
|
return 0;
|
|
|
|
err_put_source_ep_fwnode:
|
|
fwnode_handle_put(rxport->source.ep_fwnode);
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
|
|
struct fwnode_handle *link_fwnode,
|
|
struct fwnode_handle *ep_fwnode)
|
|
{
|
|
static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
|
|
"vpoc0", "vpoc1", "vpoc2", "vpoc3"
|
|
};
|
|
struct device *dev = &priv->client->dev;
|
|
struct ub960_rxport *rxport;
|
|
int ret;
|
|
|
|
rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
|
|
if (!rxport)
|
|
return -ENOMEM;
|
|
|
|
priv->rxports[nport] = rxport;
|
|
|
|
rxport->nport = nport;
|
|
rxport->priv = priv;
|
|
|
|
ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
|
|
if (ret)
|
|
goto err_free_rxport;
|
|
|
|
rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
|
|
if (IS_ERR(rxport->vpoc)) {
|
|
ret = PTR_ERR(rxport->vpoc);
|
|
if (ret == -ENODEV) {
|
|
rxport->vpoc = NULL;
|
|
} else {
|
|
dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
|
|
nport, ret);
|
|
goto err_put_remote_fwnode;
|
|
}
|
|
}
|
|
|
|
ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
|
|
if (ret)
|
|
goto err_put_remote_fwnode;
|
|
|
|
return 0;
|
|
|
|
err_put_remote_fwnode:
|
|
fwnode_handle_put(rxport->ser.fwnode);
|
|
err_free_rxport:
|
|
priv->rxports[nport] = NULL;
|
|
kfree(rxport);
|
|
return ret;
|
|
}
|
|
|
|
static struct fwnode_handle *
|
|
ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
|
|
unsigned int nport)
|
|
{
|
|
struct fwnode_handle *link_fwnode;
|
|
int ret;
|
|
|
|
fwnode_for_each_child_node(links_fwnode, link_fwnode) {
|
|
u32 link_num;
|
|
|
|
if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
|
|
continue;
|
|
|
|
ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
|
|
if (ret) {
|
|
fwnode_handle_put(link_fwnode);
|
|
return NULL;
|
|
}
|
|
|
|
if (nport == link_num)
|
|
return link_fwnode;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int ub960_parse_dt_rxports(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
struct fwnode_handle *links_fwnode;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
|
|
if (!links_fwnode) {
|
|
dev_err(dev, "'links' node missing\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
/* Defaults, recommended by TI */
|
|
priv->strobe.min = 2;
|
|
priv->strobe.max = 3;
|
|
|
|
priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct fwnode_handle *link_fwnode;
|
|
struct fwnode_handle *ep_fwnode;
|
|
|
|
link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
|
|
if (!link_fwnode)
|
|
continue;
|
|
|
|
ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
|
|
nport, 0, 0);
|
|
if (!ep_fwnode) {
|
|
fwnode_handle_put(link_fwnode);
|
|
continue;
|
|
}
|
|
|
|
ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
|
|
ep_fwnode);
|
|
|
|
fwnode_handle_put(link_fwnode);
|
|
fwnode_handle_put(ep_fwnode);
|
|
|
|
if (ret) {
|
|
dev_err(dev, "rx%u: failed to parse RX port\n", nport);
|
|
goto err_put_links;
|
|
}
|
|
}
|
|
|
|
fwnode_handle_put(links_fwnode);
|
|
|
|
return 0;
|
|
|
|
err_put_links:
|
|
fwnode_handle_put(links_fwnode);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_parse_dt_txports(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u32 nport;
|
|
int ret;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
|
|
unsigned int port = nport + priv->hw_data->num_rxports;
|
|
struct fwnode_handle *ep_fwnode;
|
|
|
|
ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
|
|
port, 0, 0);
|
|
if (!ep_fwnode)
|
|
continue;
|
|
|
|
ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
|
|
|
|
fwnode_handle_put(ep_fwnode);
|
|
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_parse_dt(struct ub960_data *priv)
|
|
{
|
|
int ret;
|
|
|
|
ret = ub960_parse_dt_rxports(priv);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = ub960_parse_dt_txports(priv);
|
|
if (ret)
|
|
goto err_free_rxports;
|
|
|
|
return 0;
|
|
|
|
err_free_rxports:
|
|
ub960_rxport_free_ports(priv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_connection *asd)
|
|
{
|
|
struct ub960_data *priv = sd_to_ub960(notifier->sd);
|
|
struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
|
|
struct device *dev = &priv->client->dev;
|
|
u8 nport = rxport->nport;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
ret = media_entity_get_fwnode_pad(&subdev->entity,
|
|
rxport->source.ep_fwnode,
|
|
MEDIA_PAD_FL_SOURCE);
|
|
if (ret < 0) {
|
|
dev_err(dev, "Failed to find pad for %s\n", subdev->name);
|
|
return ret;
|
|
}
|
|
|
|
rxport->source.sd = subdev;
|
|
rxport->source.pad = ret;
|
|
|
|
ret = media_create_pad_link(&rxport->source.sd->entity,
|
|
rxport->source.pad, &priv->sd.entity, nport,
|
|
MEDIA_LNK_FL_ENABLED |
|
|
MEDIA_LNK_FL_IMMUTABLE);
|
|
if (ret) {
|
|
dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
|
|
rxport->source.sd->name, rxport->source.pad,
|
|
priv->sd.name, nport);
|
|
return ret;
|
|
}
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
|
|
dev_dbg(dev, "Waiting for more subdevs to be bound\n");
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_connection *asd)
|
|
{
|
|
struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
|
|
|
|
rxport->source.sd = NULL;
|
|
}
|
|
|
|
static const struct v4l2_async_notifier_operations ub960_notify_ops = {
|
|
.bound = ub960_notify_bound,
|
|
.unbind = ub960_notify_unbind,
|
|
};
|
|
|
|
static int ub960_v4l2_notifier_register(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
v4l2_async_subdev_nf_init(&priv->notifier, &priv->sd);
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports; i++) {
|
|
struct ub960_rxport *rxport = priv->rxports[i];
|
|
struct ub960_asd *asd;
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
asd = v4l2_async_nf_add_fwnode(&priv->notifier,
|
|
rxport->source.ep_fwnode,
|
|
struct ub960_asd);
|
|
if (IS_ERR(asd)) {
|
|
dev_err(dev, "Failed to add subdev for source %u: %pe",
|
|
i, asd);
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
return PTR_ERR(asd);
|
|
}
|
|
|
|
asd->rxport = rxport;
|
|
}
|
|
|
|
priv->notifier.ops = &ub960_notify_ops;
|
|
|
|
ret = v4l2_async_nf_register(&priv->notifier);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to register subdev_notifier");
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
|
|
{
|
|
v4l2_async_nf_unregister(&priv->notifier);
|
|
v4l2_async_nf_cleanup(&priv->notifier);
|
|
}
|
|
|
|
static int ub960_create_subdev(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int i;
|
|
int ret;
|
|
|
|
v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
|
|
|
|
v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
|
|
priv->sd.ctrl_handler = &priv->ctrl_handler;
|
|
|
|
v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
|
|
ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
|
|
priv->tx_link_freq);
|
|
|
|
if (priv->ctrl_handler.error) {
|
|
ret = priv->ctrl_handler.error;
|
|
goto err_free_ctrl;
|
|
}
|
|
|
|
priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
|
|
V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
|
|
priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
|
|
priv->sd.entity.ops = &ub960_entity_ops;
|
|
|
|
for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
|
|
priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
|
|
MEDIA_PAD_FL_SINK :
|
|
MEDIA_PAD_FL_SOURCE;
|
|
}
|
|
|
|
ret = media_entity_pads_init(&priv->sd.entity,
|
|
priv->hw_data->num_rxports +
|
|
priv->hw_data->num_txports,
|
|
priv->pads);
|
|
if (ret)
|
|
goto err_free_ctrl;
|
|
|
|
priv->sd.state_lock = priv->sd.ctrl_handler->lock;
|
|
|
|
ret = v4l2_subdev_init_finalize(&priv->sd);
|
|
if (ret)
|
|
goto err_entity_cleanup;
|
|
|
|
ret = ub960_v4l2_notifier_register(priv);
|
|
if (ret) {
|
|
dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
|
|
goto err_subdev_cleanup;
|
|
}
|
|
|
|
ret = v4l2_async_register_subdev(&priv->sd);
|
|
if (ret) {
|
|
dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
|
|
goto err_unreg_notif;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_unreg_notif:
|
|
ub960_v4l2_notifier_unregister(priv);
|
|
err_subdev_cleanup:
|
|
v4l2_subdev_cleanup(&priv->sd);
|
|
err_entity_cleanup:
|
|
media_entity_cleanup(&priv->sd.entity);
|
|
err_free_ctrl:
|
|
v4l2_ctrl_handler_free(&priv->ctrl_handler);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_destroy_subdev(struct ub960_data *priv)
|
|
{
|
|
ub960_v4l2_notifier_unregister(priv);
|
|
v4l2_async_unregister_subdev(&priv->sd);
|
|
|
|
v4l2_subdev_cleanup(&priv->sd);
|
|
|
|
media_entity_cleanup(&priv->sd.entity);
|
|
v4l2_ctrl_handler_free(&priv->ctrl_handler);
|
|
}
|
|
|
|
static const struct regmap_config ub960_regmap_config = {
|
|
.name = "ds90ub960",
|
|
|
|
.reg_bits = 8,
|
|
.val_bits = 8,
|
|
|
|
.max_register = 0xff,
|
|
|
|
/*
|
|
* We do locking in the driver to cover the TX/RX port selection and the
|
|
* indirect register access.
|
|
*/
|
|
.disable_locking = true,
|
|
};
|
|
|
|
static void ub960_reset(struct ub960_data *priv, bool reset_regs)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
unsigned int v;
|
|
int ret;
|
|
u8 bit;
|
|
|
|
bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
|
|
UB960_SR_RESET_DIGITAL_RESET0;
|
|
|
|
ub960_write(priv, UB960_SR_RESET, bit);
|
|
|
|
mutex_lock(&priv->reg_lock);
|
|
|
|
ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
|
|
(v & bit) == 0, 2000, 100000);
|
|
|
|
mutex_unlock(&priv->reg_lock);
|
|
|
|
if (ret)
|
|
dev_err(dev, "reset failed: %d\n", ret);
|
|
}
|
|
|
|
static int ub960_get_hw_resources(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
|
|
priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
|
|
if (IS_ERR(priv->regmap))
|
|
return PTR_ERR(priv->regmap);
|
|
|
|
priv->vddio = devm_regulator_get(dev, "vddio");
|
|
if (IS_ERR(priv->vddio))
|
|
return dev_err_probe(dev, PTR_ERR(priv->vddio),
|
|
"cannot get VDDIO regulator\n");
|
|
|
|
/* get power-down pin from DT */
|
|
priv->pd_gpio =
|
|
devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
|
|
if (IS_ERR(priv->pd_gpio))
|
|
return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
|
|
"Cannot get powerdown GPIO\n");
|
|
|
|
priv->refclk = devm_clk_get(dev, "refclk");
|
|
if (IS_ERR(priv->refclk))
|
|
return dev_err_probe(dev, PTR_ERR(priv->refclk),
|
|
"Cannot get REFCLK\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ub960_enable_core_hw(struct ub960_data *priv)
|
|
{
|
|
struct device *dev = &priv->client->dev;
|
|
u8 rev_mask;
|
|
int ret;
|
|
u8 dev_sts;
|
|
u8 refclk_freq;
|
|
|
|
ret = regulator_enable(priv->vddio);
|
|
if (ret)
|
|
return dev_err_probe(dev, ret,
|
|
"failed to enable VDDIO regulator\n");
|
|
|
|
ret = clk_prepare_enable(priv->refclk);
|
|
if (ret) {
|
|
dev_err_probe(dev, ret, "Failed to enable refclk\n");
|
|
goto err_disable_vddio;
|
|
}
|
|
|
|
if (priv->pd_gpio) {
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
/* wait min 2 ms for reset to complete */
|
|
fsleep(2000);
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 0);
|
|
/* wait min 2 ms for power up to finish */
|
|
fsleep(2000);
|
|
}
|
|
|
|
ub960_reset(priv, true);
|
|
|
|
/* Runtime check register accessibility */
|
|
ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
|
|
if (ret) {
|
|
dev_err_probe(dev, ret, "Cannot read first register, abort\n");
|
|
goto err_pd_gpio;
|
|
}
|
|
|
|
dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
|
|
rev_mask);
|
|
|
|
ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
|
|
!!(dev_sts & BIT(4)), refclk_freq,
|
|
clk_get_rate(priv->refclk) / 1000000);
|
|
|
|
/* Disable all RX ports by default */
|
|
ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
|
|
/* release GPIO lock */
|
|
if (priv->hw_data->is_ub9702) {
|
|
ret = ub960_update_bits(priv, UB960_SR_RESET,
|
|
UB960_SR_RESET_GPIO_LOCK_RELEASE,
|
|
UB960_SR_RESET_GPIO_LOCK_RELEASE);
|
|
if (ret)
|
|
goto err_pd_gpio;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_pd_gpio:
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
clk_disable_unprepare(priv->refclk);
|
|
err_disable_vddio:
|
|
regulator_disable(priv->vddio);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_disable_core_hw(struct ub960_data *priv)
|
|
{
|
|
gpiod_set_value_cansleep(priv->pd_gpio, 1);
|
|
clk_disable_unprepare(priv->refclk);
|
|
regulator_disable(priv->vddio);
|
|
}
|
|
|
|
static int ub960_probe(struct i2c_client *client)
|
|
{
|
|
struct device *dev = &client->dev;
|
|
struct ub960_data *priv;
|
|
unsigned int port_lock_mask;
|
|
unsigned int port_mask;
|
|
unsigned int nport;
|
|
int ret;
|
|
|
|
priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
|
|
if (!priv)
|
|
return -ENOMEM;
|
|
|
|
priv->client = client;
|
|
|
|
priv->hw_data = device_get_match_data(dev);
|
|
|
|
mutex_init(&priv->reg_lock);
|
|
|
|
INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
|
|
|
|
/*
|
|
* Initialize these to invalid values so that the first reg writes will
|
|
* configure the target.
|
|
*/
|
|
priv->reg_current.indirect_target = 0xff;
|
|
priv->reg_current.rxport = 0xff;
|
|
priv->reg_current.txport = 0xff;
|
|
|
|
ret = ub960_get_hw_resources(priv);
|
|
if (ret)
|
|
goto err_mutex_destroy;
|
|
|
|
ret = ub960_enable_core_hw(priv);
|
|
if (ret)
|
|
goto err_mutex_destroy;
|
|
|
|
ret = ub960_parse_dt(priv);
|
|
if (ret)
|
|
goto err_disable_core_hw;
|
|
|
|
ret = ub960_init_tx_ports(priv);
|
|
if (ret)
|
|
goto err_free_ports;
|
|
|
|
ret = ub960_rxport_enable_vpocs(priv);
|
|
if (ret)
|
|
goto err_free_ports;
|
|
|
|
ret = ub960_init_rx_ports(priv);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
ub960_reset(priv, false);
|
|
|
|
port_mask = 0;
|
|
|
|
for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
|
|
struct ub960_rxport *rxport = priv->rxports[nport];
|
|
|
|
if (!rxport)
|
|
continue;
|
|
|
|
port_mask |= BIT(nport);
|
|
}
|
|
|
|
ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
if (port_mask != port_lock_mask) {
|
|
ret = -EIO;
|
|
dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
|
|
goto err_disable_vpocs;
|
|
}
|
|
|
|
/*
|
|
* Clear any errors caused by switching the RX port settings while
|
|
* probing.
|
|
*/
|
|
ub960_clear_rx_errors(priv);
|
|
|
|
ret = ub960_init_atr(priv);
|
|
if (ret)
|
|
goto err_disable_vpocs;
|
|
|
|
ret = ub960_rxport_add_serializers(priv);
|
|
if (ret)
|
|
goto err_uninit_atr;
|
|
|
|
ret = ub960_create_subdev(priv);
|
|
if (ret)
|
|
goto err_free_sers;
|
|
|
|
if (client->irq)
|
|
dev_warn(dev, "irq support not implemented, using polling\n");
|
|
|
|
schedule_delayed_work(&priv->poll_work,
|
|
msecs_to_jiffies(UB960_POLL_TIME_MS));
|
|
|
|
return 0;
|
|
|
|
err_free_sers:
|
|
ub960_rxport_remove_serializers(priv);
|
|
err_uninit_atr:
|
|
ub960_uninit_atr(priv);
|
|
err_disable_vpocs:
|
|
ub960_rxport_disable_vpocs(priv);
|
|
err_free_ports:
|
|
ub960_rxport_free_ports(priv);
|
|
ub960_txport_free_ports(priv);
|
|
err_disable_core_hw:
|
|
ub960_disable_core_hw(priv);
|
|
err_mutex_destroy:
|
|
mutex_destroy(&priv->reg_lock);
|
|
return ret;
|
|
}
|
|
|
|
static void ub960_remove(struct i2c_client *client)
|
|
{
|
|
struct v4l2_subdev *sd = i2c_get_clientdata(client);
|
|
struct ub960_data *priv = sd_to_ub960(sd);
|
|
|
|
cancel_delayed_work_sync(&priv->poll_work);
|
|
|
|
ub960_destroy_subdev(priv);
|
|
ub960_rxport_remove_serializers(priv);
|
|
ub960_uninit_atr(priv);
|
|
ub960_rxport_disable_vpocs(priv);
|
|
ub960_rxport_free_ports(priv);
|
|
ub960_txport_free_ports(priv);
|
|
ub960_disable_core_hw(priv);
|
|
mutex_destroy(&priv->reg_lock);
|
|
}
|
|
|
|
static const struct ub960_hw_data ds90ub960_hw = {
|
|
.model = "ub960",
|
|
.num_rxports = 4,
|
|
.num_txports = 2,
|
|
};
|
|
|
|
static const struct ub960_hw_data ds90ub9702_hw = {
|
|
.model = "ub9702",
|
|
.num_rxports = 4,
|
|
.num_txports = 2,
|
|
.is_ub9702 = true,
|
|
.is_fpdlink4 = true,
|
|
};
|
|
|
|
static const struct i2c_device_id ub960_id[] = {
|
|
{ "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
|
|
{ "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(i2c, ub960_id);
|
|
|
|
static const struct of_device_id ub960_dt_ids[] = {
|
|
{ .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
|
|
{ .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(of, ub960_dt_ids);
|
|
|
|
static struct i2c_driver ds90ub960_driver = {
|
|
.probe = ub960_probe,
|
|
.remove = ub960_remove,
|
|
.id_table = ub960_id,
|
|
.driver = {
|
|
.name = "ds90ub960",
|
|
.of_match_table = ub960_dt_ids,
|
|
},
|
|
};
|
|
module_i2c_driver(ds90ub960_driver);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
|
|
MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
|
|
MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
|
|
MODULE_IMPORT_NS(I2C_ATR);
|