linux-stable/drivers/iio/adc/ti-ads131e08.c
Linus Torvalds 4a0225c3d2 spi: Updates for v5.13
The only core work for SPI this time around is the completion of the
 conversion to the new style method for specifying transfer delays,
 meaning we can cope with what most controllers support more directly
 using conversions in the core rather than open coding in drivers.
 Otherwise it's a good stack of cleanups and fixes plus a few new
 drivers.
 
 The conversion to new style transfer delay will cause an issue with a
 newly added staging driver which has a straightforward resolution in
 -next.
 
  - Completion of the conversion to new style transfer delay
    configuration.
  - Introduction and use of module_parport_driver() helper, merged here
    as there's no parport tree.
  - Support for Altera SoCs on DFL buses, NXP i.MX8DL, HiSilicon Kunpeng,
    MediaTek MT8195,
 -----BEGIN PGP SIGNATURE-----
 
 iQEzBAABCgAdFiEEreZoqmdXGLWf4p/qJNaLcl1Uh9AFAmCG0FYACgkQJNaLcl1U
 h9BXlAf/ZQaU8Nq4NY2jgjzjUoplF4qSFvCZ05CXqfYftaAOp06AmoOwNuNTz2gU
 Fkxouuw3D0hNmaeVusF1PaRjIvJCu+RO68EDrzTJlcFytyC8CxWSDE4Yw6ytOBUM
 OWZdVXsuw0CUk3VRJl2ycCooeTyKaCksfkVucocZAoyexcfQrFpxkOCUbA8hVM43
 Hghzb8HWAZUerrfwreSwfvyVMralR3rqqbZFKgSgT/sRM3zpaR4sctIkNgKKEbFE
 eRPRfBIWWZdJtjQ+uifFAc3jJHeZlmNXuQq3C+ETd2vQDFlymTxj+U3u74ieZxrm
 c6V4u4R3+Qx9qONV/q0LV6E4sBSvdw==
 =cBUy
 -----END PGP SIGNATURE-----

Merge tag 'spi-v5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi

Pull spi updates from Mark Brown:
 "The only core work for SPI this time around is the completion of the
  conversion to the new style method for specifying transfer delays,
  meaning we can cope with what most controllers support more directly
  using conversions in the core rather than open coding in drivers.

  Otherwise it's a good stack of cleanups and fixes plus a few new
  drivers.

  Summary:

   - Completion of the conversion to new style transfer delay
     configuration

   - Introduction and use of module_parport_driver() helper, merged here
     as there's no parport tree

   - Support for Altera SoCs on DFL buses, NXP i.MX8DL, HiSilicon
     Kunpeng, MediaTek MT8195"

* tag 'spi-v5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi: (113 commits)
  spi: Rename enable1 to activate in spi_set_cs()
  spi: Convert Freescale QSPI binding to json schema
  spi: stm32-qspi: fix debug format string
  spi: tools: make a symbolic link to the header file spi.h
  spi: fsi: add a missing of_node_put
  spi: Make error handling of gpiod_count() call cleaner
  spidev: Add Micron SPI NOR Authenta device compatible
  spi: brcm,spi-bcm-qspi: convert to the json-schema
  spi: altera: Add DFL bus driver for Altera API Controller
  spi: altera: separate core code from platform code
  spi: stm32-qspi: Fix compilation warning in ARM64
  spi: Handle SPI device setup callback failure.
  spi: sync up initial chipselect state
  spi: stm32-qspi: Add dirmap support
  spi: stm32-qspi: Trigger DMA only if more than 4 bytes to transfer
  spi: stm32-qspi: fix pm_runtime usage_count counter
  spi: spi-zynqmp-gqspi: return -ENOMEM if dma_map_single fails
  spi: spi-zynqmp-gqspi: fix use-after-free in zynqmp_qspi_exec_op
  spi: spi-zynqmp-gqspi: Resolved slab-out-of-bounds bug
  spi: spi-zynqmp-gqspi: fix hang issue when suspend/resume
  ...
2021-04-26 16:32:11 -07:00

954 lines
23 KiB
C

// SPDX-License-Identifier: GPL-2.0
/*
* Texas Instruments ADS131E0x 4-, 6- and 8-Channel ADCs
*
* Copyright (c) 2020 AVL DiTEST GmbH
* Tomislav Denis <tomislav.denis@avl.com>
*
* Datasheet: https://www.ti.com/lit/ds/symlink/ads131e08.pdf
*/
#include <linux/bitfield.h>
#include <linux/clk.h>
#include <linux/delay.h>
#include <linux/module.h>
#include <linux/iio/buffer.h>
#include <linux/iio/iio.h>
#include <linux/iio/sysfs.h>
#include <linux/iio/trigger.h>
#include <linux/iio/trigger_consumer.h>
#include <linux/iio/triggered_buffer.h>
#include <linux/regulator/consumer.h>
#include <linux/spi/spi.h>
#include <asm/unaligned.h>
/* Commands */
#define ADS131E08_CMD_RESET 0x06
#define ADS131E08_CMD_START 0x08
#define ADS131E08_CMD_STOP 0x0A
#define ADS131E08_CMD_OFFSETCAL 0x1A
#define ADS131E08_CMD_SDATAC 0x11
#define ADS131E08_CMD_RDATA 0x12
#define ADS131E08_CMD_RREG(r) (BIT(5) | (r & GENMASK(4, 0)))
#define ADS131E08_CMD_WREG(r) (BIT(6) | (r & GENMASK(4, 0)))
/* Registers */
#define ADS131E08_ADR_CFG1R 0x01
#define ADS131E08_ADR_CFG3R 0x03
#define ADS131E08_ADR_CH0R 0x05
/* Configuration register 1 */
#define ADS131E08_CFG1R_DR_MASK GENMASK(2, 0)
/* Configuration register 3 */
#define ADS131E08_CFG3R_PDB_REFBUF_MASK BIT(7)
#define ADS131E08_CFG3R_VREF_4V_MASK BIT(5)
/* Channel settings register */
#define ADS131E08_CHR_GAIN_MASK GENMASK(6, 4)
#define ADS131E08_CHR_MUX_MASK GENMASK(2, 0)
#define ADS131E08_CHR_PWD_MASK BIT(7)
/* ADC misc */
#define ADS131E08_DEFAULT_DATA_RATE 1
#define ADS131E08_DEFAULT_PGA_GAIN 1
#define ADS131E08_DEFAULT_MUX 0
#define ADS131E08_VREF_2V4_mV 2400
#define ADS131E08_VREF_4V_mV 4000
#define ADS131E08_WAIT_RESET_CYCLES 18
#define ADS131E08_WAIT_SDECODE_CYCLES 4
#define ADS131E08_WAIT_OFFSETCAL_MS 153
#define ADS131E08_MAX_SETTLING_TIME_MS 6
#define ADS131E08_NUM_STATUS_BYTES 3
#define ADS131E08_NUM_DATA_BYTES_MAX 24
#define ADS131E08_NUM_DATA_BYTES(dr) (((dr) >= 32) ? 2 : 3)
#define ADS131E08_NUM_DATA_BITS(dr) (ADS131E08_NUM_DATA_BYTES(dr) * 8)
#define ADS131E08_NUM_STORAGE_BYTES 4
enum ads131e08_ids {
ads131e04,
ads131e06,
ads131e08,
};
struct ads131e08_info {
unsigned int max_channels;
const char *name;
};
struct ads131e08_channel_config {
unsigned int pga_gain;
unsigned int mux;
};
struct ads131e08_state {
const struct ads131e08_info *info;
struct spi_device *spi;
struct iio_trigger *trig;
struct clk *adc_clk;
struct regulator *vref_reg;
struct ads131e08_channel_config *channel_config;
unsigned int data_rate;
unsigned int vref_mv;
unsigned int sdecode_delay_us;
unsigned int reset_delay_us;
unsigned int readback_len;
struct completion completion;
struct {
u8 data[ADS131E08_NUM_DATA_BYTES_MAX];
s64 ts __aligned(8);
} tmp_buf;
u8 tx_buf[3] ____cacheline_aligned;
/*
* Add extra one padding byte to be able to access the last channel
* value using u32 pointer
*/
u8 rx_buf[ADS131E08_NUM_STATUS_BYTES +
ADS131E08_NUM_DATA_BYTES_MAX + 1];
};
static const struct ads131e08_info ads131e08_info_tbl[] = {
[ads131e04] = {
.max_channels = 4,
.name = "ads131e04",
},
[ads131e06] = {
.max_channels = 6,
.name = "ads131e06",
},
[ads131e08] = {
.max_channels = 8,
.name = "ads131e08",
},
};
struct ads131e08_data_rate_desc {
unsigned int rate; /* data rate in kSPS */
u8 reg; /* reg value */
};
static const struct ads131e08_data_rate_desc ads131e08_data_rate_tbl[] = {
{ .rate = 64, .reg = 0x00 },
{ .rate = 32, .reg = 0x01 },
{ .rate = 16, .reg = 0x02 },
{ .rate = 8, .reg = 0x03 },
{ .rate = 4, .reg = 0x04 },
{ .rate = 2, .reg = 0x05 },
{ .rate = 1, .reg = 0x06 },
};
struct ads131e08_pga_gain_desc {
unsigned int gain; /* PGA gain value */
u8 reg; /* field value */
};
static const struct ads131e08_pga_gain_desc ads131e08_pga_gain_tbl[] = {
{ .gain = 1, .reg = 0x01 },
{ .gain = 2, .reg = 0x02 },
{ .gain = 4, .reg = 0x04 },
{ .gain = 8, .reg = 0x05 },
{ .gain = 12, .reg = 0x06 },
};
static const u8 ads131e08_valid_channel_mux_values[] = { 0, 1, 3, 4 };
static int ads131e08_exec_cmd(struct ads131e08_state *st, u8 cmd)
{
int ret;
ret = spi_write_then_read(st->spi, &cmd, 1, NULL, 0);
if (ret)
dev_err(&st->spi->dev, "Exec cmd(%02x) failed\n", cmd);
return ret;
}
static int ads131e08_read_reg(struct ads131e08_state *st, u8 reg)
{
int ret;
struct spi_transfer transfer[] = {
{
.tx_buf = &st->tx_buf,
.len = 2,
.delay = {
.value = st->sdecode_delay_us,
.unit = SPI_DELAY_UNIT_USECS,
},
}, {
.rx_buf = &st->rx_buf,
.len = 1,
},
};
st->tx_buf[0] = ADS131E08_CMD_RREG(reg);
st->tx_buf[1] = 0;
ret = spi_sync_transfer(st->spi, transfer, ARRAY_SIZE(transfer));
if (ret) {
dev_err(&st->spi->dev, "Read register failed\n");
return ret;
}
return st->rx_buf[0];
}
static int ads131e08_write_reg(struct ads131e08_state *st, u8 reg, u8 value)
{
int ret;
struct spi_transfer transfer[] = {
{
.tx_buf = &st->tx_buf,
.len = 3,
.delay = {
.value = st->sdecode_delay_us,
.unit = SPI_DELAY_UNIT_USECS,
},
}
};
st->tx_buf[0] = ADS131E08_CMD_WREG(reg);
st->tx_buf[1] = 0;
st->tx_buf[2] = value;
ret = spi_sync_transfer(st->spi, transfer, ARRAY_SIZE(transfer));
if (ret)
dev_err(&st->spi->dev, "Write register failed\n");
return ret;
}
static int ads131e08_read_data(struct ads131e08_state *st, int rx_len)
{
int ret;
struct spi_transfer transfer[] = {
{
.tx_buf = &st->tx_buf,
.len = 1,
}, {
.rx_buf = &st->rx_buf,
.len = rx_len,
},
};
st->tx_buf[0] = ADS131E08_CMD_RDATA;
ret = spi_sync_transfer(st->spi, transfer, ARRAY_SIZE(transfer));
if (ret)
dev_err(&st->spi->dev, "Read data failed\n");
return ret;
}
static int ads131e08_set_data_rate(struct ads131e08_state *st, int data_rate)
{
int i, reg, ret;
for (i = 0; i < ARRAY_SIZE(ads131e08_data_rate_tbl); i++) {
if (ads131e08_data_rate_tbl[i].rate == data_rate)
break;
}
if (i == ARRAY_SIZE(ads131e08_data_rate_tbl)) {
dev_err(&st->spi->dev, "invalid data rate value\n");
return -EINVAL;
}
reg = ads131e08_read_reg(st, ADS131E08_ADR_CFG1R);
if (reg < 0)
return reg;
reg &= ~ADS131E08_CFG1R_DR_MASK;
reg |= FIELD_PREP(ADS131E08_CFG1R_DR_MASK,
ads131e08_data_rate_tbl[i].reg);
ret = ads131e08_write_reg(st, ADS131E08_ADR_CFG1R, reg);
if (ret)
return ret;
st->data_rate = data_rate;
st->readback_len = ADS131E08_NUM_STATUS_BYTES +
ADS131E08_NUM_DATA_BYTES(st->data_rate) *
st->info->max_channels;
return 0;
}
static int ads131e08_pga_gain_to_field_value(struct ads131e08_state *st,
unsigned int pga_gain)
{
int i;
for (i = 0; i < ARRAY_SIZE(ads131e08_pga_gain_tbl); i++) {
if (ads131e08_pga_gain_tbl[i].gain == pga_gain)
break;
}
if (i == ARRAY_SIZE(ads131e08_pga_gain_tbl)) {
dev_err(&st->spi->dev, "invalid PGA gain value\n");
return -EINVAL;
}
return ads131e08_pga_gain_tbl[i].reg;
}
static int ads131e08_set_pga_gain(struct ads131e08_state *st,
unsigned int channel, unsigned int pga_gain)
{
int field_value, reg;
field_value = ads131e08_pga_gain_to_field_value(st, pga_gain);
if (field_value < 0)
return field_value;
reg = ads131e08_read_reg(st, ADS131E08_ADR_CH0R + channel);
if (reg < 0)
return reg;
reg &= ~ADS131E08_CHR_GAIN_MASK;
reg |= FIELD_PREP(ADS131E08_CHR_GAIN_MASK, field_value);
return ads131e08_write_reg(st, ADS131E08_ADR_CH0R + channel, reg);
}
static int ads131e08_validate_channel_mux(struct ads131e08_state *st,
unsigned int mux)
{
int i;
for (i = 0; i < ARRAY_SIZE(ads131e08_valid_channel_mux_values); i++) {
if (ads131e08_valid_channel_mux_values[i] == mux)
break;
}
if (i == ARRAY_SIZE(ads131e08_valid_channel_mux_values)) {
dev_err(&st->spi->dev, "invalid channel mux value\n");
return -EINVAL;
}
return 0;
}
static int ads131e08_set_channel_mux(struct ads131e08_state *st,
unsigned int channel, unsigned int mux)
{
int reg;
reg = ads131e08_read_reg(st, ADS131E08_ADR_CH0R + channel);
if (reg < 0)
return reg;
reg &= ~ADS131E08_CHR_MUX_MASK;
reg |= FIELD_PREP(ADS131E08_CHR_MUX_MASK, mux);
return ads131e08_write_reg(st, ADS131E08_ADR_CH0R + channel, reg);
}
static int ads131e08_power_down_channel(struct ads131e08_state *st,
unsigned int channel, bool value)
{
int reg;
reg = ads131e08_read_reg(st, ADS131E08_ADR_CH0R + channel);
if (reg < 0)
return reg;
reg &= ~ADS131E08_CHR_PWD_MASK;
reg |= FIELD_PREP(ADS131E08_CHR_PWD_MASK, value);
return ads131e08_write_reg(st, ADS131E08_ADR_CH0R + channel, reg);
}
static int ads131e08_config_reference_voltage(struct ads131e08_state *st)
{
int reg;
reg = ads131e08_read_reg(st, ADS131E08_ADR_CFG3R);
if (reg < 0)
return reg;
reg &= ~ADS131E08_CFG3R_PDB_REFBUF_MASK;
if (!st->vref_reg) {
reg |= FIELD_PREP(ADS131E08_CFG3R_PDB_REFBUF_MASK, 1);
reg &= ~ADS131E08_CFG3R_VREF_4V_MASK;
reg |= FIELD_PREP(ADS131E08_CFG3R_VREF_4V_MASK,
st->vref_mv == ADS131E08_VREF_4V_mV);
}
return ads131e08_write_reg(st, ADS131E08_ADR_CFG3R, reg);
}
static int ads131e08_initial_config(struct iio_dev *indio_dev)
{
const struct iio_chan_spec *channel = indio_dev->channels;
struct ads131e08_state *st = iio_priv(indio_dev);
unsigned long active_channels = 0;
int ret, i;
ret = ads131e08_exec_cmd(st, ADS131E08_CMD_RESET);
if (ret)
return ret;
udelay(st->reset_delay_us);
/* Disable read data in continuous mode (enabled by default) */
ret = ads131e08_exec_cmd(st, ADS131E08_CMD_SDATAC);
if (ret)
return ret;
ret = ads131e08_set_data_rate(st, ADS131E08_DEFAULT_DATA_RATE);
if (ret)
return ret;
ret = ads131e08_config_reference_voltage(st);
if (ret)
return ret;
for (i = 0; i < indio_dev->num_channels; i++) {
ret = ads131e08_set_pga_gain(st, channel->channel,
st->channel_config[i].pga_gain);
if (ret)
return ret;
ret = ads131e08_set_channel_mux(st, channel->channel,
st->channel_config[i].mux);
if (ret)
return ret;
active_channels |= BIT(channel->channel);
channel++;
}
/* Power down unused channels */
for_each_clear_bit(i, &active_channels, st->info->max_channels) {
ret = ads131e08_power_down_channel(st, i, true);
if (ret)
return ret;
}
/* Request channel offset calibration */
ret = ads131e08_exec_cmd(st, ADS131E08_CMD_OFFSETCAL);
if (ret)
return ret;
/*
* Channel offset calibration is triggered with the first START
* command. Since calibration takes more time than settling operation,
* this causes timeout error when command START is sent first
* time (e.g. first call of the ads131e08_read_direct method).
* To avoid this problem offset calibration is triggered here.
*/
ret = ads131e08_exec_cmd(st, ADS131E08_CMD_START);
if (ret)
return ret;
msleep(ADS131E08_WAIT_OFFSETCAL_MS);
return ads131e08_exec_cmd(st, ADS131E08_CMD_STOP);
}
static int ads131e08_pool_data(struct ads131e08_state *st)
{
unsigned long timeout;
int ret;
reinit_completion(&st->completion);
ret = ads131e08_exec_cmd(st, ADS131E08_CMD_START);
if (ret)
return ret;
timeout = msecs_to_jiffies(ADS131E08_MAX_SETTLING_TIME_MS);
ret = wait_for_completion_timeout(&st->completion, timeout);
if (!ret)
return -ETIMEDOUT;
ret = ads131e08_read_data(st, st->readback_len);
if (ret)
return ret;
return ads131e08_exec_cmd(st, ADS131E08_CMD_STOP);
}
static int ads131e08_read_direct(struct iio_dev *indio_dev,
struct iio_chan_spec const *channel, int *value)
{
struct ads131e08_state *st = iio_priv(indio_dev);
u8 num_bits, *src;
int ret;
ret = ads131e08_pool_data(st);
if (ret)
return ret;
src = st->rx_buf + ADS131E08_NUM_STATUS_BYTES +
channel->channel * ADS131E08_NUM_DATA_BYTES(st->data_rate);
num_bits = ADS131E08_NUM_DATA_BITS(st->data_rate);
*value = sign_extend32(get_unaligned_be32(src) >> (32 - num_bits), num_bits - 1);
return 0;
}
static int ads131e08_read_raw(struct iio_dev *indio_dev,
struct iio_chan_spec const *channel, int *value,
int *value2, long mask)
{
struct ads131e08_state *st = iio_priv(indio_dev);
int ret;
switch (mask) {
case IIO_CHAN_INFO_RAW:
ret = iio_device_claim_direct_mode(indio_dev);
if (ret)
return ret;
ret = ads131e08_read_direct(indio_dev, channel, value);
iio_device_release_direct_mode(indio_dev);
if (ret)
return ret;
return IIO_VAL_INT;
case IIO_CHAN_INFO_SCALE:
if (st->vref_reg) {
ret = regulator_get_voltage(st->vref_reg);
if (ret < 0)
return ret;
*value = ret / 1000;
} else {
*value = st->vref_mv;
}
*value /= st->channel_config[channel->address].pga_gain;
*value2 = ADS131E08_NUM_DATA_BITS(st->data_rate) - 1;
return IIO_VAL_FRACTIONAL_LOG2;
case IIO_CHAN_INFO_SAMP_FREQ:
*value = st->data_rate;
return IIO_VAL_INT;
default:
return -EINVAL;
}
}
static int ads131e08_write_raw(struct iio_dev *indio_dev,
struct iio_chan_spec const *channel, int value,
int value2, long mask)
{
struct ads131e08_state *st = iio_priv(indio_dev);
int ret;
switch (mask) {
case IIO_CHAN_INFO_SAMP_FREQ:
ret = iio_device_claim_direct_mode(indio_dev);
if (ret)
return ret;
ret = ads131e08_set_data_rate(st, value);
iio_device_release_direct_mode(indio_dev);
return ret;
default:
return -EINVAL;
}
}
static IIO_CONST_ATTR_SAMP_FREQ_AVAIL("1 2 4 8 16 32 64");
static struct attribute *ads131e08_attributes[] = {
&iio_const_attr_sampling_frequency_available.dev_attr.attr,
NULL
};
static const struct attribute_group ads131e08_attribute_group = {
.attrs = ads131e08_attributes,
};
static int ads131e08_debugfs_reg_access(struct iio_dev *indio_dev,
unsigned int reg, unsigned int writeval, unsigned int *readval)
{
struct ads131e08_state *st = iio_priv(indio_dev);
if (readval) {
int ret = ads131e08_read_reg(st, reg);
*readval = ret;
return ret;
}
return ads131e08_write_reg(st, reg, writeval);
}
static const struct iio_info ads131e08_iio_info = {
.read_raw = ads131e08_read_raw,
.write_raw = ads131e08_write_raw,
.attrs = &ads131e08_attribute_group,
.debugfs_reg_access = &ads131e08_debugfs_reg_access,
};
static int ads131e08_set_trigger_state(struct iio_trigger *trig, bool state)
{
struct iio_dev *indio_dev = iio_trigger_get_drvdata(trig);
struct ads131e08_state *st = iio_priv(indio_dev);
u8 cmd = state ? ADS131E08_CMD_START : ADS131E08_CMD_STOP;
return ads131e08_exec_cmd(st, cmd);
}
static const struct iio_trigger_ops ads131e08_trigger_ops = {
.set_trigger_state = &ads131e08_set_trigger_state,
.validate_device = &iio_trigger_validate_own_device,
};
static irqreturn_t ads131e08_trigger_handler(int irq, void *private)
{
struct iio_poll_func *pf = private;
struct iio_dev *indio_dev = pf->indio_dev;
struct ads131e08_state *st = iio_priv(indio_dev);
unsigned int chn, i = 0;
u8 *src, *dest;
int ret;
/*
* The number of data bits per channel depends on the data rate.
* For 32 and 64 ksps data rates, number of data bits per channel
* is 16. This case is not compliant with used (fixed) scan element
* type (be:s24/32>>8). So we use a little tweak to pack properly
* 16 bits of data into the buffer.
*/
unsigned int num_bytes = ADS131E08_NUM_DATA_BYTES(st->data_rate);
u8 tweek_offset = num_bytes == 2 ? 1 : 0;
if (iio_trigger_using_own(indio_dev))
ret = ads131e08_read_data(st, st->readback_len);
else
ret = ads131e08_pool_data(st);
if (ret)
goto out;
for_each_set_bit(chn, indio_dev->active_scan_mask, indio_dev->masklength) {
src = st->rx_buf + ADS131E08_NUM_STATUS_BYTES + chn * num_bytes;
dest = st->tmp_buf.data + i * ADS131E08_NUM_STORAGE_BYTES;
/*
* Tweek offset is 0:
* +---+---+---+---+
* |D0 |D1 |D2 | X | (3 data bytes)
* +---+---+---+---+
* a+0 a+1 a+2 a+3
*
* Tweek offset is 1:
* +---+---+---+---+
* |P0 |D0 |D1 | X | (one padding byte and 2 data bytes)
* +---+---+---+---+
* a+0 a+1 a+2 a+3
*/
memcpy(dest + tweek_offset, src, num_bytes);
/*
* Data conversion from 16 bits of data to 24 bits of data
* is done by sign extension (properly filling padding byte).
*/
if (tweek_offset)
*dest = *src & BIT(7) ? 0xff : 0x00;
i++;
}
iio_push_to_buffers_with_timestamp(indio_dev, st->tmp_buf.data,
iio_get_time_ns(indio_dev));
out:
iio_trigger_notify_done(indio_dev->trig);
return IRQ_HANDLED;
}
static irqreturn_t ads131e08_interrupt(int irq, void *private)
{
struct iio_dev *indio_dev = private;
struct ads131e08_state *st = iio_priv(indio_dev);
if (iio_buffer_enabled(indio_dev) && iio_trigger_using_own(indio_dev))
iio_trigger_poll(st->trig);
else
complete(&st->completion);
return IRQ_HANDLED;
}
static int ads131e08_alloc_channels(struct iio_dev *indio_dev)
{
struct ads131e08_state *st = iio_priv(indio_dev);
struct ads131e08_channel_config *channel_config;
struct device *dev = &st->spi->dev;
struct iio_chan_spec *channels;
struct fwnode_handle *node;
unsigned int channel, tmp;
int num_channels, i, ret;
ret = device_property_read_u32(dev, "ti,vref-internal", &tmp);
if (ret)
tmp = 0;
switch (tmp) {
case 0:
st->vref_mv = ADS131E08_VREF_2V4_mV;
break;
case 1:
st->vref_mv = ADS131E08_VREF_4V_mV;
break;
default:
dev_err(&st->spi->dev, "invalid internal voltage reference\n");
return -EINVAL;
}
num_channels = device_get_child_node_count(dev);
if (num_channels == 0) {
dev_err(&st->spi->dev, "no channel children\n");
return -ENODEV;
}
if (num_channels > st->info->max_channels) {
dev_err(&st->spi->dev, "num of channel children out of range\n");
return -EINVAL;
}
channels = devm_kcalloc(&st->spi->dev, num_channels,
sizeof(*channels), GFP_KERNEL);
if (!channels)
return -ENOMEM;
channel_config = devm_kcalloc(&st->spi->dev, num_channels,
sizeof(*channel_config), GFP_KERNEL);
if (!channel_config)
return -ENOMEM;
i = 0;
device_for_each_child_node(dev, node) {
ret = fwnode_property_read_u32(node, "reg", &channel);
if (ret)
return ret;
ret = fwnode_property_read_u32(node, "ti,gain", &tmp);
if (ret) {
channel_config[i].pga_gain = ADS131E08_DEFAULT_PGA_GAIN;
} else {
ret = ads131e08_pga_gain_to_field_value(st, tmp);
if (ret < 0)
return ret;
channel_config[i].pga_gain = tmp;
}
ret = fwnode_property_read_u32(node, "ti,mux", &tmp);
if (ret) {
channel_config[i].mux = ADS131E08_DEFAULT_MUX;
} else {
ret = ads131e08_validate_channel_mux(st, tmp);
if (ret)
return ret;
channel_config[i].mux = tmp;
}
channels[i].type = IIO_VOLTAGE;
channels[i].indexed = 1;
channels[i].channel = channel;
channels[i].address = i;
channels[i].info_mask_separate = BIT(IIO_CHAN_INFO_RAW) |
BIT(IIO_CHAN_INFO_SCALE);
channels[i].info_mask_shared_by_type = BIT(IIO_CHAN_INFO_SAMP_FREQ);
channels[i].scan_index = channel;
channels[i].scan_type.sign = 's';
channels[i].scan_type.realbits = 24;
channels[i].scan_type.storagebits = 32;
channels[i].scan_type.shift = 8;
channels[i].scan_type.endianness = IIO_BE;
i++;
}
indio_dev->channels = channels;
indio_dev->num_channels = num_channels;
st->channel_config = channel_config;
return 0;
}
static void ads131e08_regulator_disable(void *data)
{
struct ads131e08_state *st = data;
regulator_disable(st->vref_reg);
}
static void ads131e08_clk_disable(void *data)
{
struct ads131e08_state *st = data;
clk_disable_unprepare(st->adc_clk);
}
static int ads131e08_probe(struct spi_device *spi)
{
const struct ads131e08_info *info;
struct ads131e08_state *st;
struct iio_dev *indio_dev;
unsigned long adc_clk_hz;
unsigned long adc_clk_ns;
int ret;
info = device_get_match_data(&spi->dev);
if (!info) {
dev_err(&spi->dev, "failed to get match data\n");
return -ENODEV;
}
indio_dev = devm_iio_device_alloc(&spi->dev, sizeof(*st));
if (!indio_dev) {
dev_err(&spi->dev, "failed to allocate IIO device\n");
return -ENOMEM;
}
st = iio_priv(indio_dev);
st->info = info;
st->spi = spi;
ret = ads131e08_alloc_channels(indio_dev);
if (ret)
return ret;
indio_dev->name = st->info->name;
indio_dev->dev.parent = &spi->dev;
indio_dev->info = &ads131e08_iio_info;
indio_dev->modes = INDIO_DIRECT_MODE;
init_completion(&st->completion);
if (spi->irq) {
ret = devm_request_irq(&spi->dev, spi->irq,
ads131e08_interrupt,
IRQF_TRIGGER_FALLING | IRQF_ONESHOT,
spi->dev.driver->name, indio_dev);
if (ret)
return dev_err_probe(&spi->dev, ret,
"request irq failed\n");
} else {
dev_err(&spi->dev, "data ready IRQ missing\n");
return -ENODEV;
}
st->trig = devm_iio_trigger_alloc(&spi->dev, "%s-dev%d",
indio_dev->name, indio_dev->id);
if (!st->trig) {
dev_err(&spi->dev, "failed to allocate IIO trigger\n");
return -ENOMEM;
}
st->trig->ops = &ads131e08_trigger_ops;
st->trig->dev.parent = &spi->dev;
iio_trigger_set_drvdata(st->trig, indio_dev);
ret = devm_iio_trigger_register(&spi->dev, st->trig);
if (ret) {
dev_err(&spi->dev, "failed to register IIO trigger\n");
return -ENOMEM;
}
indio_dev->trig = iio_trigger_get(st->trig);
ret = devm_iio_triggered_buffer_setup(&spi->dev, indio_dev,
NULL, &ads131e08_trigger_handler, NULL);
if (ret) {
dev_err(&spi->dev, "failed to setup IIO buffer\n");
return ret;
}
st->vref_reg = devm_regulator_get_optional(&spi->dev, "vref");
if (!IS_ERR(st->vref_reg)) {
ret = regulator_enable(st->vref_reg);
if (ret) {
dev_err(&spi->dev,
"failed to enable external vref supply\n");
return ret;
}
ret = devm_add_action_or_reset(&spi->dev, ads131e08_regulator_disable, st);
if (ret)
return ret;
} else {
if (PTR_ERR(st->vref_reg) != -ENODEV)
return PTR_ERR(st->vref_reg);
st->vref_reg = NULL;
}
st->adc_clk = devm_clk_get(&spi->dev, "adc-clk");
if (IS_ERR(st->adc_clk))
return dev_err_probe(&spi->dev, PTR_ERR(st->adc_clk),
"failed to get the ADC clock\n");
ret = clk_prepare_enable(st->adc_clk);
if (ret) {
dev_err(&spi->dev, "failed to prepare/enable the ADC clock\n");
return ret;
}
ret = devm_add_action_or_reset(&spi->dev, ads131e08_clk_disable, st);
if (ret)
return ret;
adc_clk_hz = clk_get_rate(st->adc_clk);
if (!adc_clk_hz) {
dev_err(&spi->dev, "failed to get the ADC clock rate\n");
return -EINVAL;
}
adc_clk_ns = NSEC_PER_SEC / adc_clk_hz;
st->sdecode_delay_us = DIV_ROUND_UP(
ADS131E08_WAIT_SDECODE_CYCLES * adc_clk_ns, NSEC_PER_USEC);
st->reset_delay_us = DIV_ROUND_UP(
ADS131E08_WAIT_RESET_CYCLES * adc_clk_ns, NSEC_PER_USEC);
ret = ads131e08_initial_config(indio_dev);
if (ret) {
dev_err(&spi->dev, "initial configuration failed\n");
return ret;
}
return devm_iio_device_register(&spi->dev, indio_dev);
}
static const struct of_device_id ads131e08_of_match[] = {
{ .compatible = "ti,ads131e04",
.data = &ads131e08_info_tbl[ads131e04], },
{ .compatible = "ti,ads131e06",
.data = &ads131e08_info_tbl[ads131e06], },
{ .compatible = "ti,ads131e08",
.data = &ads131e08_info_tbl[ads131e08], },
{}
};
MODULE_DEVICE_TABLE(of, ads131e08_of_match);
static struct spi_driver ads131e08_driver = {
.driver = {
.name = "ads131e08",
.of_match_table = ads131e08_of_match,
},
.probe = ads131e08_probe,
};
module_spi_driver(ads131e08_driver);
MODULE_AUTHOR("Tomislav Denis <tomislav.denis@avl.com>");
MODULE_DESCRIPTION("Driver for ADS131E0x ADC family");
MODULE_LICENSE("GPL v2");