mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-15 15:04:27 +08:00
66500bbc7d
do __le32_to_cpu to tlv data fields to make sure driver runs correctly when running on big endian system. Signed-off-by: Kiran K <kiran.k@intel.com> Reviewed-by: Chethan T N <chethan.tumkur.narayan@intel.com> Reviewed-by: Srivatsa Ravishankar <ravishankar.srivatsa@intel.com> Signed-off-by: Marcel Holtmann <marcel@holtmann.org> Signed-off-by: Johan Hedberg <johan.hedberg@intel.com>
1110 lines
27 KiB
C
1110 lines
27 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
*
|
|
* Bluetooth support for Intel devices
|
|
*
|
|
* Copyright (C) 2015 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/firmware.h>
|
|
#include <linux/regmap.h>
|
|
#include <asm/unaligned.h>
|
|
|
|
#include <net/bluetooth/bluetooth.h>
|
|
#include <net/bluetooth/hci_core.h>
|
|
|
|
#include "btintel.h"
|
|
|
|
#define VERSION "0.1"
|
|
|
|
#define BDADDR_INTEL (&(bdaddr_t){{0x00, 0x8b, 0x9e, 0x19, 0x03, 0x00}})
|
|
#define RSA_HEADER_LEN 644
|
|
#define CSS_HEADER_OFFSET 8
|
|
#define ECDSA_OFFSET 644
|
|
#define ECDSA_HEADER_LEN 320
|
|
|
|
int btintel_check_bdaddr(struct hci_dev *hdev)
|
|
{
|
|
struct hci_rp_read_bd_addr *bda;
|
|
struct sk_buff *skb;
|
|
|
|
skb = __hci_cmd_sync(hdev, HCI_OP_READ_BD_ADDR, 0, NULL,
|
|
HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
int err = PTR_ERR(skb);
|
|
bt_dev_err(hdev, "Reading Intel device address failed (%d)",
|
|
err);
|
|
return err;
|
|
}
|
|
|
|
if (skb->len != sizeof(*bda)) {
|
|
bt_dev_err(hdev, "Intel device address length mismatch");
|
|
kfree_skb(skb);
|
|
return -EIO;
|
|
}
|
|
|
|
bda = (struct hci_rp_read_bd_addr *)skb->data;
|
|
|
|
/* For some Intel based controllers, the default Bluetooth device
|
|
* address 00:03:19:9E:8B:00 can be found. These controllers are
|
|
* fully operational, but have the danger of duplicate addresses
|
|
* and that in turn can cause problems with Bluetooth operation.
|
|
*/
|
|
if (!bacmp(&bda->bdaddr, BDADDR_INTEL)) {
|
|
bt_dev_err(hdev, "Found Intel default device address (%pMR)",
|
|
&bda->bdaddr);
|
|
set_bit(HCI_QUIRK_INVALID_BDADDR, &hdev->quirks);
|
|
}
|
|
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_check_bdaddr);
|
|
|
|
int btintel_enter_mfg(struct hci_dev *hdev)
|
|
{
|
|
static const u8 param[] = { 0x01, 0x00 };
|
|
struct sk_buff *skb;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc11, 2, param, HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Entering manufacturer mode failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_enter_mfg);
|
|
|
|
int btintel_exit_mfg(struct hci_dev *hdev, bool reset, bool patched)
|
|
{
|
|
u8 param[] = { 0x00, 0x00 };
|
|
struct sk_buff *skb;
|
|
|
|
/* The 2nd command parameter specifies the manufacturing exit method:
|
|
* 0x00: Just disable the manufacturing mode (0x00).
|
|
* 0x01: Disable manufacturing mode and reset with patches deactivated.
|
|
* 0x02: Disable manufacturing mode and reset with patches activated.
|
|
*/
|
|
if (reset)
|
|
param[1] |= patched ? 0x02 : 0x01;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc11, 2, param, HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Exiting manufacturer mode failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_exit_mfg);
|
|
|
|
int btintel_set_bdaddr(struct hci_dev *hdev, const bdaddr_t *bdaddr)
|
|
{
|
|
struct sk_buff *skb;
|
|
int err;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc31, 6, bdaddr, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
err = PTR_ERR(skb);
|
|
bt_dev_err(hdev, "Changing Intel device address failed (%d)",
|
|
err);
|
|
return err;
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_bdaddr);
|
|
|
|
int btintel_set_diag(struct hci_dev *hdev, bool enable)
|
|
{
|
|
struct sk_buff *skb;
|
|
u8 param[3];
|
|
int err;
|
|
|
|
if (enable) {
|
|
param[0] = 0x03;
|
|
param[1] = 0x03;
|
|
param[2] = 0x03;
|
|
} else {
|
|
param[0] = 0x00;
|
|
param[1] = 0x00;
|
|
param[2] = 0x00;
|
|
}
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc43, 3, param, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
err = PTR_ERR(skb);
|
|
if (err == -ENODATA)
|
|
goto done;
|
|
bt_dev_err(hdev, "Changing Intel diagnostic mode failed (%d)",
|
|
err);
|
|
return err;
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
done:
|
|
btintel_set_event_mask(hdev, enable);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_diag);
|
|
|
|
int btintel_set_diag_mfg(struct hci_dev *hdev, bool enable)
|
|
{
|
|
int err, ret;
|
|
|
|
err = btintel_enter_mfg(hdev);
|
|
if (err)
|
|
return err;
|
|
|
|
ret = btintel_set_diag(hdev, enable);
|
|
|
|
err = btintel_exit_mfg(hdev, false, false);
|
|
if (err)
|
|
return err;
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_diag_mfg);
|
|
|
|
void btintel_hw_error(struct hci_dev *hdev, u8 code)
|
|
{
|
|
struct sk_buff *skb;
|
|
u8 type = 0x00;
|
|
|
|
bt_dev_err(hdev, "Hardware error 0x%2.2x", code);
|
|
|
|
skb = __hci_cmd_sync(hdev, HCI_OP_RESET, 0, NULL, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Reset after hardware error failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return;
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc22, 1, &type, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Retrieving Intel exception info failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return;
|
|
}
|
|
|
|
if (skb->len != 13) {
|
|
bt_dev_err(hdev, "Exception info size mismatch");
|
|
kfree_skb(skb);
|
|
return;
|
|
}
|
|
|
|
bt_dev_err(hdev, "Exception info %s", (char *)(skb->data + 1));
|
|
|
|
kfree_skb(skb);
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_hw_error);
|
|
|
|
void btintel_version_info(struct hci_dev *hdev, struct intel_version *ver)
|
|
{
|
|
const char *variant;
|
|
|
|
switch (ver->fw_variant) {
|
|
case 0x06:
|
|
variant = "Bootloader";
|
|
break;
|
|
case 0x23:
|
|
variant = "Firmware";
|
|
break;
|
|
default:
|
|
return;
|
|
}
|
|
|
|
bt_dev_info(hdev, "%s revision %u.%u build %u week %u %u",
|
|
variant, ver->fw_revision >> 4, ver->fw_revision & 0x0f,
|
|
ver->fw_build_num, ver->fw_build_ww,
|
|
2000 + ver->fw_build_yy);
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_version_info);
|
|
|
|
int btintel_secure_send(struct hci_dev *hdev, u8 fragment_type, u32 plen,
|
|
const void *param)
|
|
{
|
|
while (plen > 0) {
|
|
struct sk_buff *skb;
|
|
u8 cmd_param[253], fragment_len = (plen > 252) ? 252 : plen;
|
|
|
|
cmd_param[0] = fragment_type;
|
|
memcpy(cmd_param + 1, param, fragment_len);
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc09, fragment_len + 1,
|
|
cmd_param, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb))
|
|
return PTR_ERR(skb);
|
|
|
|
kfree_skb(skb);
|
|
|
|
plen -= fragment_len;
|
|
param += fragment_len;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_secure_send);
|
|
|
|
int btintel_load_ddc_config(struct hci_dev *hdev, const char *ddc_name)
|
|
{
|
|
const struct firmware *fw;
|
|
struct sk_buff *skb;
|
|
const u8 *fw_ptr;
|
|
int err;
|
|
|
|
err = request_firmware_direct(&fw, ddc_name, &hdev->dev);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to load Intel DDC file %s (%d)",
|
|
ddc_name, err);
|
|
return err;
|
|
}
|
|
|
|
bt_dev_info(hdev, "Found Intel DDC parameters: %s", ddc_name);
|
|
|
|
fw_ptr = fw->data;
|
|
|
|
/* DDC file contains one or more DDC structure which has
|
|
* Length (1 byte), DDC ID (2 bytes), and DDC value (Length - 2).
|
|
*/
|
|
while (fw->size > fw_ptr - fw->data) {
|
|
u8 cmd_plen = fw_ptr[0] + sizeof(u8);
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc8b, cmd_plen, fw_ptr,
|
|
HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Failed to send Intel_Write_DDC (%ld)",
|
|
PTR_ERR(skb));
|
|
release_firmware(fw);
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
fw_ptr += cmd_plen;
|
|
kfree_skb(skb);
|
|
}
|
|
|
|
release_firmware(fw);
|
|
|
|
bt_dev_info(hdev, "Applying Intel DDC parameters completed");
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_load_ddc_config);
|
|
|
|
int btintel_set_event_mask(struct hci_dev *hdev, bool debug)
|
|
{
|
|
u8 mask[8] = { 0x87, 0x0c, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 };
|
|
struct sk_buff *skb;
|
|
int err;
|
|
|
|
if (debug)
|
|
mask[1] |= 0x62;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc52, 8, mask, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
err = PTR_ERR(skb);
|
|
bt_dev_err(hdev, "Setting Intel event mask failed (%d)", err);
|
|
return err;
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_event_mask);
|
|
|
|
int btintel_set_event_mask_mfg(struct hci_dev *hdev, bool debug)
|
|
{
|
|
int err, ret;
|
|
|
|
err = btintel_enter_mfg(hdev);
|
|
if (err)
|
|
return err;
|
|
|
|
ret = btintel_set_event_mask(hdev, debug);
|
|
|
|
err = btintel_exit_mfg(hdev, false, false);
|
|
if (err)
|
|
return err;
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_event_mask_mfg);
|
|
|
|
int btintel_read_version(struct hci_dev *hdev, struct intel_version *ver)
|
|
{
|
|
struct sk_buff *skb;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc05, 0, NULL, HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Reading Intel version information failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
if (skb->len != sizeof(*ver)) {
|
|
bt_dev_err(hdev, "Intel version event size mismatch");
|
|
kfree_skb(skb);
|
|
return -EILSEQ;
|
|
}
|
|
|
|
memcpy(ver, skb->data, sizeof(*ver));
|
|
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_read_version);
|
|
|
|
void btintel_version_info_tlv(struct hci_dev *hdev, struct intel_version_tlv *version)
|
|
{
|
|
const char *variant;
|
|
|
|
switch (version->img_type) {
|
|
case 0x01:
|
|
variant = "Bootloader";
|
|
bt_dev_info(hdev, "Device revision is %u", version->dev_rev_id);
|
|
bt_dev_info(hdev, "Secure boot is %s",
|
|
version->secure_boot ? "enabled" : "disabled");
|
|
bt_dev_info(hdev, "OTP lock is %s",
|
|
version->otp_lock ? "enabled" : "disabled");
|
|
bt_dev_info(hdev, "API lock is %s",
|
|
version->api_lock ? "enabled" : "disabled");
|
|
bt_dev_info(hdev, "Debug lock is %s",
|
|
version->debug_lock ? "enabled" : "disabled");
|
|
bt_dev_info(hdev, "Minimum firmware build %u week %u %u",
|
|
version->min_fw_build_nn, version->min_fw_build_cw,
|
|
2000 + version->min_fw_build_yy);
|
|
break;
|
|
case 0x03:
|
|
variant = "Firmware";
|
|
break;
|
|
default:
|
|
bt_dev_err(hdev, "Unsupported image type(%02x)", version->img_type);
|
|
goto done;
|
|
}
|
|
|
|
bt_dev_info(hdev, "%s timestamp %u.%u buildtype %u build %u", variant,
|
|
2000 + (version->timestamp >> 8), version->timestamp & 0xff,
|
|
version->build_type, version->build_num);
|
|
|
|
done:
|
|
return;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_version_info_tlv);
|
|
|
|
int btintel_read_version_tlv(struct hci_dev *hdev, struct intel_version_tlv *version)
|
|
{
|
|
struct sk_buff *skb;
|
|
const u8 param[1] = { 0xFF };
|
|
|
|
if (!version)
|
|
return -EINVAL;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc05, 1, param, HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Reading Intel version information failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
if (skb->data[0]) {
|
|
bt_dev_err(hdev, "Intel Read Version command failed (%02x)",
|
|
skb->data[0]);
|
|
kfree_skb(skb);
|
|
return -EIO;
|
|
}
|
|
|
|
/* Consume Command Complete Status field */
|
|
skb_pull(skb, 1);
|
|
|
|
/* Event parameters contatin multiple TLVs. Read each of them
|
|
* and only keep the required data. Also, it use existing legacy
|
|
* version field like hw_platform, hw_variant, and fw_variant
|
|
* to keep the existing setup flow
|
|
*/
|
|
while (skb->len) {
|
|
struct intel_tlv *tlv;
|
|
|
|
tlv = (struct intel_tlv *)skb->data;
|
|
switch (tlv->type) {
|
|
case INTEL_TLV_CNVI_TOP:
|
|
version->cnvi_top =
|
|
__le32_to_cpu(get_unaligned_le32(tlv->val));
|
|
break;
|
|
case INTEL_TLV_CNVR_TOP:
|
|
version->cnvr_top =
|
|
__le32_to_cpu(get_unaligned_le32(tlv->val));
|
|
break;
|
|
case INTEL_TLV_CNVI_BT:
|
|
version->cnvi_bt =
|
|
__le32_to_cpu(get_unaligned_le32(tlv->val));
|
|
break;
|
|
case INTEL_TLV_CNVR_BT:
|
|
version->cnvr_bt =
|
|
__le32_to_cpu(get_unaligned_le32(tlv->val));
|
|
break;
|
|
case INTEL_TLV_DEV_REV_ID:
|
|
version->dev_rev_id =
|
|
__le16_to_cpu(get_unaligned_le16(tlv->val));
|
|
break;
|
|
case INTEL_TLV_IMAGE_TYPE:
|
|
version->img_type = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_TIME_STAMP:
|
|
version->timestamp =
|
|
__le16_to_cpu(get_unaligned_le16(tlv->val));
|
|
break;
|
|
case INTEL_TLV_BUILD_TYPE:
|
|
version->build_type = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_BUILD_NUM:
|
|
version->build_num =
|
|
__le32_to_cpu(get_unaligned_le32(tlv->val));
|
|
break;
|
|
case INTEL_TLV_SECURE_BOOT:
|
|
version->secure_boot = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_OTP_LOCK:
|
|
version->otp_lock = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_API_LOCK:
|
|
version->api_lock = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_DEBUG_LOCK:
|
|
version->debug_lock = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_MIN_FW:
|
|
version->min_fw_build_nn = tlv->val[0];
|
|
version->min_fw_build_cw = tlv->val[1];
|
|
version->min_fw_build_yy = tlv->val[2];
|
|
break;
|
|
case INTEL_TLV_LIMITED_CCE:
|
|
version->limited_cce = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_SBE_TYPE:
|
|
version->sbe_type = tlv->val[0];
|
|
break;
|
|
case INTEL_TLV_OTP_BDADDR:
|
|
memcpy(&version->otp_bd_addr, tlv->val, tlv->len);
|
|
break;
|
|
default:
|
|
/* Ignore rest of information */
|
|
break;
|
|
}
|
|
/* consume the current tlv and move to next*/
|
|
skb_pull(skb, tlv->len + sizeof(*tlv));
|
|
}
|
|
|
|
kfree_skb(skb);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_read_version_tlv);
|
|
|
|
/* ------- REGMAP IBT SUPPORT ------- */
|
|
|
|
#define IBT_REG_MODE_8BIT 0x00
|
|
#define IBT_REG_MODE_16BIT 0x01
|
|
#define IBT_REG_MODE_32BIT 0x02
|
|
|
|
struct regmap_ibt_context {
|
|
struct hci_dev *hdev;
|
|
__u16 op_write;
|
|
__u16 op_read;
|
|
};
|
|
|
|
struct ibt_cp_reg_access {
|
|
__le32 addr;
|
|
__u8 mode;
|
|
__u8 len;
|
|
__u8 data[];
|
|
} __packed;
|
|
|
|
struct ibt_rp_reg_access {
|
|
__u8 status;
|
|
__le32 addr;
|
|
__u8 data[];
|
|
} __packed;
|
|
|
|
static int regmap_ibt_read(void *context, const void *addr, size_t reg_size,
|
|
void *val, size_t val_size)
|
|
{
|
|
struct regmap_ibt_context *ctx = context;
|
|
struct ibt_cp_reg_access cp;
|
|
struct ibt_rp_reg_access *rp;
|
|
struct sk_buff *skb;
|
|
int err = 0;
|
|
|
|
if (reg_size != sizeof(__le32))
|
|
return -EINVAL;
|
|
|
|
switch (val_size) {
|
|
case 1:
|
|
cp.mode = IBT_REG_MODE_8BIT;
|
|
break;
|
|
case 2:
|
|
cp.mode = IBT_REG_MODE_16BIT;
|
|
break;
|
|
case 4:
|
|
cp.mode = IBT_REG_MODE_32BIT;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* regmap provides a little-endian formatted addr */
|
|
cp.addr = *(__le32 *)addr;
|
|
cp.len = val_size;
|
|
|
|
bt_dev_dbg(ctx->hdev, "Register (0x%x) read", le32_to_cpu(cp.addr));
|
|
|
|
skb = hci_cmd_sync(ctx->hdev, ctx->op_read, sizeof(cp), &cp,
|
|
HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
err = PTR_ERR(skb);
|
|
bt_dev_err(ctx->hdev, "regmap: Register (0x%x) read error (%d)",
|
|
le32_to_cpu(cp.addr), err);
|
|
return err;
|
|
}
|
|
|
|
if (skb->len != sizeof(*rp) + val_size) {
|
|
bt_dev_err(ctx->hdev, "regmap: Register (0x%x) read error, bad len",
|
|
le32_to_cpu(cp.addr));
|
|
err = -EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
rp = (struct ibt_rp_reg_access *)skb->data;
|
|
|
|
if (rp->addr != cp.addr) {
|
|
bt_dev_err(ctx->hdev, "regmap: Register (0x%x) read error, bad addr",
|
|
le32_to_cpu(rp->addr));
|
|
err = -EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
memcpy(val, rp->data, val_size);
|
|
|
|
done:
|
|
kfree_skb(skb);
|
|
return err;
|
|
}
|
|
|
|
static int regmap_ibt_gather_write(void *context,
|
|
const void *addr, size_t reg_size,
|
|
const void *val, size_t val_size)
|
|
{
|
|
struct regmap_ibt_context *ctx = context;
|
|
struct ibt_cp_reg_access *cp;
|
|
struct sk_buff *skb;
|
|
int plen = sizeof(*cp) + val_size;
|
|
u8 mode;
|
|
int err = 0;
|
|
|
|
if (reg_size != sizeof(__le32))
|
|
return -EINVAL;
|
|
|
|
switch (val_size) {
|
|
case 1:
|
|
mode = IBT_REG_MODE_8BIT;
|
|
break;
|
|
case 2:
|
|
mode = IBT_REG_MODE_16BIT;
|
|
break;
|
|
case 4:
|
|
mode = IBT_REG_MODE_32BIT;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
cp = kmalloc(plen, GFP_KERNEL);
|
|
if (!cp)
|
|
return -ENOMEM;
|
|
|
|
/* regmap provides a little-endian formatted addr/value */
|
|
cp->addr = *(__le32 *)addr;
|
|
cp->mode = mode;
|
|
cp->len = val_size;
|
|
memcpy(&cp->data, val, val_size);
|
|
|
|
bt_dev_dbg(ctx->hdev, "Register (0x%x) write", le32_to_cpu(cp->addr));
|
|
|
|
skb = hci_cmd_sync(ctx->hdev, ctx->op_write, plen, cp, HCI_CMD_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
err = PTR_ERR(skb);
|
|
bt_dev_err(ctx->hdev, "regmap: Register (0x%x) write error (%d)",
|
|
le32_to_cpu(cp->addr), err);
|
|
goto done;
|
|
}
|
|
kfree_skb(skb);
|
|
|
|
done:
|
|
kfree(cp);
|
|
return err;
|
|
}
|
|
|
|
static int regmap_ibt_write(void *context, const void *data, size_t count)
|
|
{
|
|
/* data contains register+value, since we only support 32bit addr,
|
|
* minimum data size is 4 bytes.
|
|
*/
|
|
if (WARN_ONCE(count < 4, "Invalid register access"))
|
|
return -EINVAL;
|
|
|
|
return regmap_ibt_gather_write(context, data, 4, data + 4, count - 4);
|
|
}
|
|
|
|
static void regmap_ibt_free_context(void *context)
|
|
{
|
|
kfree(context);
|
|
}
|
|
|
|
static struct regmap_bus regmap_ibt = {
|
|
.read = regmap_ibt_read,
|
|
.write = regmap_ibt_write,
|
|
.gather_write = regmap_ibt_gather_write,
|
|
.free_context = regmap_ibt_free_context,
|
|
.reg_format_endian_default = REGMAP_ENDIAN_LITTLE,
|
|
.val_format_endian_default = REGMAP_ENDIAN_LITTLE,
|
|
};
|
|
|
|
/* Config is the same for all register regions */
|
|
static const struct regmap_config regmap_ibt_cfg = {
|
|
.name = "btintel_regmap",
|
|
.reg_bits = 32,
|
|
.val_bits = 32,
|
|
};
|
|
|
|
struct regmap *btintel_regmap_init(struct hci_dev *hdev, u16 opcode_read,
|
|
u16 opcode_write)
|
|
{
|
|
struct regmap_ibt_context *ctx;
|
|
|
|
bt_dev_info(hdev, "regmap: Init R%x-W%x region", opcode_read,
|
|
opcode_write);
|
|
|
|
ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
|
|
if (!ctx)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
ctx->op_read = opcode_read;
|
|
ctx->op_write = opcode_write;
|
|
ctx->hdev = hdev;
|
|
|
|
return regmap_init(&hdev->dev, ®map_ibt, ctx, ®map_ibt_cfg);
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_regmap_init);
|
|
|
|
int btintel_send_intel_reset(struct hci_dev *hdev, u32 boot_param)
|
|
{
|
|
struct intel_reset params = { 0x00, 0x01, 0x00, 0x01, 0x00000000 };
|
|
struct sk_buff *skb;
|
|
|
|
params.boot_param = cpu_to_le32(boot_param);
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc01, sizeof(params), ¶ms,
|
|
HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Failed to send Intel Reset command");
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
kfree_skb(skb);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_send_intel_reset);
|
|
|
|
int btintel_read_boot_params(struct hci_dev *hdev,
|
|
struct intel_boot_params *params)
|
|
{
|
|
struct sk_buff *skb;
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc0d, 0, NULL, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Reading Intel boot parameters failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
if (skb->len != sizeof(*params)) {
|
|
bt_dev_err(hdev, "Intel boot parameters size mismatch");
|
|
kfree_skb(skb);
|
|
return -EILSEQ;
|
|
}
|
|
|
|
memcpy(params, skb->data, sizeof(*params));
|
|
|
|
kfree_skb(skb);
|
|
|
|
if (params->status) {
|
|
bt_dev_err(hdev, "Intel boot parameters command failed (%02x)",
|
|
params->status);
|
|
return -bt_to_errno(params->status);
|
|
}
|
|
|
|
bt_dev_info(hdev, "Device revision is %u",
|
|
le16_to_cpu(params->dev_revid));
|
|
|
|
bt_dev_info(hdev, "Secure boot is %s",
|
|
params->secure_boot ? "enabled" : "disabled");
|
|
|
|
bt_dev_info(hdev, "OTP lock is %s",
|
|
params->otp_lock ? "enabled" : "disabled");
|
|
|
|
bt_dev_info(hdev, "API lock is %s",
|
|
params->api_lock ? "enabled" : "disabled");
|
|
|
|
bt_dev_info(hdev, "Debug lock is %s",
|
|
params->debug_lock ? "enabled" : "disabled");
|
|
|
|
bt_dev_info(hdev, "Minimum firmware build %u week %u %u",
|
|
params->min_fw_build_nn, params->min_fw_build_cw,
|
|
2000 + params->min_fw_build_yy);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_read_boot_params);
|
|
|
|
static int btintel_sfi_rsa_header_secure_send(struct hci_dev *hdev,
|
|
const struct firmware *fw)
|
|
{
|
|
int err;
|
|
|
|
/* Start the firmware download transaction with the Init fragment
|
|
* represented by the 128 bytes of CSS header.
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x00, 128, fw->data);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware header (%d)", err);
|
|
goto done;
|
|
}
|
|
|
|
/* Send the 256 bytes of public key information from the firmware
|
|
* as the PKey fragment.
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x03, 256, fw->data + 128);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware pkey (%d)", err);
|
|
goto done;
|
|
}
|
|
|
|
/* Send the 256 bytes of signature information from the firmware
|
|
* as the Sign fragment.
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x02, 256, fw->data + 388);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware signature (%d)", err);
|
|
goto done;
|
|
}
|
|
|
|
done:
|
|
return err;
|
|
}
|
|
|
|
static int btintel_sfi_ecdsa_header_secure_send(struct hci_dev *hdev,
|
|
const struct firmware *fw)
|
|
{
|
|
int err;
|
|
|
|
/* Start the firmware download transaction with the Init fragment
|
|
* represented by the 128 bytes of CSS header.
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x00, 128, fw->data + 644);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware header (%d)", err);
|
|
return err;
|
|
}
|
|
|
|
/* Send the 96 bytes of public key information from the firmware
|
|
* as the PKey fragment.
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x03, 96, fw->data + 644 + 128);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware pkey (%d)", err);
|
|
return err;
|
|
}
|
|
|
|
/* Send the 96 bytes of signature information from the firmware
|
|
* as the Sign fragment
|
|
*/
|
|
err = btintel_secure_send(hdev, 0x02, 96, fw->data + 644 + 224);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev, "Failed to send firmware signature (%d)",
|
|
err);
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int btintel_download_firmware_payload(struct hci_dev *hdev,
|
|
const struct firmware *fw,
|
|
u32 *boot_param, size_t offset)
|
|
{
|
|
int err;
|
|
const u8 *fw_ptr;
|
|
u32 frag_len;
|
|
|
|
fw_ptr = fw->data + offset;
|
|
frag_len = 0;
|
|
err = -EINVAL;
|
|
|
|
while (fw_ptr - fw->data < fw->size) {
|
|
struct hci_command_hdr *cmd = (void *)(fw_ptr + frag_len);
|
|
|
|
/* Each SKU has a different reset parameter to use in the
|
|
* HCI_Intel_Reset command and it is embedded in the firmware
|
|
* data. So, instead of using static value per SKU, check
|
|
* the firmware data and save it for later use.
|
|
*/
|
|
if (le16_to_cpu(cmd->opcode) == 0xfc0e) {
|
|
/* The boot parameter is the first 32-bit value
|
|
* and rest of 3 octets are reserved.
|
|
*/
|
|
*boot_param = get_unaligned_le32(fw_ptr + sizeof(*cmd));
|
|
|
|
bt_dev_dbg(hdev, "boot_param=0x%x", *boot_param);
|
|
}
|
|
|
|
frag_len += sizeof(*cmd) + cmd->plen;
|
|
|
|
/* The parameter length of the secure send command requires
|
|
* a 4 byte alignment. It happens so that the firmware file
|
|
* contains proper Intel_NOP commands to align the fragments
|
|
* as needed.
|
|
*
|
|
* Send set of commands with 4 byte alignment from the
|
|
* firmware data buffer as a single Data fragement.
|
|
*/
|
|
if (!(frag_len % 4)) {
|
|
err = btintel_secure_send(hdev, 0x01, frag_len, fw_ptr);
|
|
if (err < 0) {
|
|
bt_dev_err(hdev,
|
|
"Failed to send firmware data (%d)",
|
|
err);
|
|
goto done;
|
|
}
|
|
|
|
fw_ptr += frag_len;
|
|
frag_len = 0;
|
|
}
|
|
}
|
|
|
|
done:
|
|
return err;
|
|
}
|
|
|
|
int btintel_download_firmware(struct hci_dev *hdev,
|
|
const struct firmware *fw,
|
|
u32 *boot_param)
|
|
{
|
|
int err;
|
|
|
|
err = btintel_sfi_rsa_header_secure_send(hdev, fw);
|
|
if (err)
|
|
return err;
|
|
|
|
return btintel_download_firmware_payload(hdev, fw, boot_param,
|
|
RSA_HEADER_LEN);
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_download_firmware);
|
|
|
|
int btintel_download_firmware_newgen(struct hci_dev *hdev,
|
|
const struct firmware *fw, u32 *boot_param,
|
|
u8 hw_variant, u8 sbe_type)
|
|
{
|
|
int err;
|
|
u32 css_header_ver;
|
|
|
|
/* iBT hardware variants 0x0b, 0x0c, 0x11, 0x12, 0x13, 0x14 support
|
|
* only RSA secure boot engine. Hence, the corresponding sfi file will
|
|
* have RSA header of 644 bytes followed by Command Buffer.
|
|
*
|
|
* iBT hardware variants 0x17, 0x18 onwards support both RSA and ECDSA
|
|
* secure boot engine. As a result, the corresponding sfi file will
|
|
* have RSA header of 644, ECDSA header of 320 bytes followed by
|
|
* Command Buffer.
|
|
*
|
|
* CSS Header byte positions 0x08 to 0x0B represent the CSS Header
|
|
* version: RSA(0x00010000) , ECDSA (0x00020000)
|
|
*/
|
|
css_header_ver = get_unaligned_le32(fw->data + CSS_HEADER_OFFSET);
|
|
if (css_header_ver != 0x00010000) {
|
|
bt_dev_err(hdev, "Invalid CSS Header version");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (hw_variant <= 0x14) {
|
|
if (sbe_type != 0x00) {
|
|
bt_dev_err(hdev, "Invalid SBE type for hardware variant (%d)",
|
|
hw_variant);
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = btintel_sfi_rsa_header_secure_send(hdev, fw);
|
|
if (err)
|
|
return err;
|
|
|
|
err = btintel_download_firmware_payload(hdev, fw, boot_param, RSA_HEADER_LEN);
|
|
if (err)
|
|
return err;
|
|
} else if (hw_variant >= 0x17) {
|
|
/* Check if CSS header for ECDSA follows the RSA header */
|
|
if (fw->data[ECDSA_OFFSET] != 0x06)
|
|
return -EINVAL;
|
|
|
|
/* Check if the CSS Header version is ECDSA(0x00020000) */
|
|
css_header_ver = get_unaligned_le32(fw->data + ECDSA_OFFSET + CSS_HEADER_OFFSET);
|
|
if (css_header_ver != 0x00020000) {
|
|
bt_dev_err(hdev, "Invalid CSS Header version");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (sbe_type == 0x00) {
|
|
err = btintel_sfi_rsa_header_secure_send(hdev, fw);
|
|
if (err)
|
|
return err;
|
|
|
|
err = btintel_download_firmware_payload(hdev, fw,
|
|
boot_param,
|
|
RSA_HEADER_LEN + ECDSA_HEADER_LEN);
|
|
if (err)
|
|
return err;
|
|
} else if (sbe_type == 0x01) {
|
|
err = btintel_sfi_ecdsa_header_secure_send(hdev, fw);
|
|
if (err)
|
|
return err;
|
|
|
|
err = btintel_download_firmware_payload(hdev, fw,
|
|
boot_param,
|
|
RSA_HEADER_LEN + ECDSA_HEADER_LEN);
|
|
if (err)
|
|
return err;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_download_firmware_newgen);
|
|
|
|
void btintel_reset_to_bootloader(struct hci_dev *hdev)
|
|
{
|
|
struct intel_reset params;
|
|
struct sk_buff *skb;
|
|
|
|
/* Send Intel Reset command. This will result in
|
|
* re-enumeration of BT controller.
|
|
*
|
|
* Intel Reset parameter description:
|
|
* reset_type : 0x00 (Soft reset),
|
|
* 0x01 (Hard reset)
|
|
* patch_enable : 0x00 (Do not enable),
|
|
* 0x01 (Enable)
|
|
* ddc_reload : 0x00 (Do not reload),
|
|
* 0x01 (Reload)
|
|
* boot_option: 0x00 (Current image),
|
|
* 0x01 (Specified boot address)
|
|
* boot_param: Boot address
|
|
*
|
|
*/
|
|
params.reset_type = 0x01;
|
|
params.patch_enable = 0x01;
|
|
params.ddc_reload = 0x01;
|
|
params.boot_option = 0x00;
|
|
params.boot_param = cpu_to_le32(0x00000000);
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc01, sizeof(params),
|
|
¶ms, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "FW download error recovery failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return;
|
|
}
|
|
bt_dev_info(hdev, "Intel reset sent to retry FW download");
|
|
kfree_skb(skb);
|
|
|
|
/* Current Intel BT controllers(ThP/JfP) hold the USB reset
|
|
* lines for 2ms when it receives Intel Reset in bootloader mode.
|
|
* Whereas, the upcoming Intel BT controllers will hold USB reset
|
|
* for 150ms. To keep the delay generic, 150ms is chosen here.
|
|
*/
|
|
msleep(150);
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_reset_to_bootloader);
|
|
|
|
int btintel_read_debug_features(struct hci_dev *hdev,
|
|
struct intel_debug_features *features)
|
|
{
|
|
struct sk_buff *skb;
|
|
u8 page_no = 1;
|
|
|
|
/* Intel controller supports two pages, each page is of 128-bit
|
|
* feature bit mask. And each bit defines specific feature support
|
|
*/
|
|
skb = __hci_cmd_sync(hdev, 0xfca6, sizeof(page_no), &page_no,
|
|
HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Reading supported features failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
if (skb->len != (sizeof(features->page1) + 3)) {
|
|
bt_dev_err(hdev, "Supported features event size mismatch");
|
|
kfree_skb(skb);
|
|
return -EILSEQ;
|
|
}
|
|
|
|
memcpy(features->page1, skb->data + 3, sizeof(features->page1));
|
|
|
|
/* Read the supported features page2 if required in future.
|
|
*/
|
|
kfree_skb(skb);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_read_debug_features);
|
|
|
|
int btintel_set_debug_features(struct hci_dev *hdev,
|
|
const struct intel_debug_features *features)
|
|
{
|
|
u8 mask[11] = { 0x0a, 0x92, 0x02, 0x07, 0x00, 0x00, 0x00, 0x00,
|
|
0x00, 0x00, 0x00 };
|
|
struct sk_buff *skb;
|
|
|
|
if (!features)
|
|
return -EINVAL;
|
|
|
|
if (!(features->page1[0] & 0x3f)) {
|
|
bt_dev_info(hdev, "Telemetry exception format not supported");
|
|
return 0;
|
|
}
|
|
|
|
skb = __hci_cmd_sync(hdev, 0xfc8b, 11, mask, HCI_INIT_TIMEOUT);
|
|
if (IS_ERR(skb)) {
|
|
bt_dev_err(hdev, "Setting Intel telemetry ddc write event mask failed (%ld)",
|
|
PTR_ERR(skb));
|
|
return PTR_ERR(skb);
|
|
}
|
|
|
|
kfree_skb(skb);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(btintel_set_debug_features);
|
|
|
|
MODULE_AUTHOR("Marcel Holtmann <marcel@holtmann.org>");
|
|
MODULE_DESCRIPTION("Bluetooth support for Intel devices ver " VERSION);
|
|
MODULE_VERSION(VERSION);
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_FIRMWARE("intel/ibt-11-5.sfi");
|
|
MODULE_FIRMWARE("intel/ibt-11-5.ddc");
|
|
MODULE_FIRMWARE("intel/ibt-12-16.sfi");
|
|
MODULE_FIRMWARE("intel/ibt-12-16.ddc");
|