1 // SPDX-License-Identifier: GPL-2.0-only
3 * The industrial I/O core
5 * Copyright (c) 2008 Jonathan Cameron
7 * Based on elements of hwmon and input subsystems.
10 #define pr_fmt(fmt) "iio-core: " fmt
12 #include <linux/anon_inodes.h>
13 #include <linux/cdev.h>
14 #include <linux/cleanup.h>
15 #include <linux/debugfs.h>
16 #include <linux/device.h>
17 #include <linux/err.h>
19 #include <linux/idr.h>
20 #include <linux/kdev_t.h>
21 #include <linux/kernel.h>
22 #include <linux/module.h>
23 #include <linux/mutex.h>
24 #include <linux/poll.h>
25 #include <linux/property.h>
26 #include <linux/sched.h>
27 #include <linux/slab.h>
28 #include <linux/wait.h>
30 #include <linux/iio/buffer.h>
31 #include <linux/iio/buffer_impl.h>
32 #include <linux/iio/events.h>
33 #include <linux/iio/iio-opaque.h>
34 #include <linux/iio/iio.h>
35 #include <linux/iio/sysfs.h>
38 #include "iio_core_trigger.h"
40 /* IDA to assign each registered device a unique id */
41 static DEFINE_IDA(iio_ida
);
43 static dev_t iio_devt
;
45 #define IIO_DEV_MAX 256
46 const struct bus_type iio_bus_type
= {
49 EXPORT_SYMBOL(iio_bus_type
);
51 static struct dentry
*iio_debugfs_dentry
;
53 static const char * const iio_direction
[] = {
58 static const char * const iio_chan_type_name_spec
[] = {
59 [IIO_VOLTAGE
] = "voltage",
60 [IIO_CURRENT
] = "current",
61 [IIO_POWER
] = "power",
62 [IIO_ACCEL
] = "accel",
63 [IIO_ANGL_VEL
] = "anglvel",
65 [IIO_LIGHT
] = "illuminance",
66 [IIO_INTENSITY
] = "intensity",
67 [IIO_PROXIMITY
] = "proximity",
69 [IIO_INCLI
] = "incli",
72 [IIO_TIMESTAMP
] = "timestamp",
73 [IIO_CAPACITANCE
] = "capacitance",
74 [IIO_ALTVOLTAGE
] = "altvoltage",
76 [IIO_PRESSURE
] = "pressure",
77 [IIO_HUMIDITYRELATIVE
] = "humidityrelative",
78 [IIO_ACTIVITY
] = "activity",
79 [IIO_STEPS
] = "steps",
80 [IIO_ENERGY
] = "energy",
81 [IIO_DISTANCE
] = "distance",
82 [IIO_VELOCITY
] = "velocity",
83 [IIO_CONCENTRATION
] = "concentration",
84 [IIO_RESISTANCE
] = "resistance",
86 [IIO_UVINDEX
] = "uvindex",
87 [IIO_ELECTRICALCONDUCTIVITY
] = "electricalconductivity",
88 [IIO_COUNT
] = "count",
89 [IIO_INDEX
] = "index",
90 [IIO_GRAVITY
] = "gravity",
91 [IIO_POSITIONRELATIVE
] = "positionrelative",
92 [IIO_PHASE
] = "phase",
93 [IIO_MASSCONCENTRATION
] = "massconcentration",
94 [IIO_DELTA_ANGL
] = "deltaangl",
95 [IIO_DELTA_VELOCITY
] = "deltavelocity",
96 [IIO_COLORTEMP
] = "colortemp",
97 [IIO_CHROMATICITY
] = "chromaticity",
98 [IIO_ATTENTION
] = "attention",
101 static const char * const iio_modifier_names
[] = {
105 [IIO_MOD_X_AND_Y
] = "x&y",
106 [IIO_MOD_X_AND_Z
] = "x&z",
107 [IIO_MOD_Y_AND_Z
] = "y&z",
108 [IIO_MOD_X_AND_Y_AND_Z
] = "x&y&z",
109 [IIO_MOD_X_OR_Y
] = "x|y",
110 [IIO_MOD_X_OR_Z
] = "x|z",
111 [IIO_MOD_Y_OR_Z
] = "y|z",
112 [IIO_MOD_X_OR_Y_OR_Z
] = "x|y|z",
113 [IIO_MOD_ROOT_SUM_SQUARED_X_Y
] = "sqrt(x^2+y^2)",
114 [IIO_MOD_SUM_SQUARED_X_Y_Z
] = "x^2+y^2+z^2",
115 [IIO_MOD_LIGHT_BOTH
] = "both",
116 [IIO_MOD_LIGHT_IR
] = "ir",
117 [IIO_MOD_LIGHT_CLEAR
] = "clear",
118 [IIO_MOD_LIGHT_RED
] = "red",
119 [IIO_MOD_LIGHT_GREEN
] = "green",
120 [IIO_MOD_LIGHT_BLUE
] = "blue",
121 [IIO_MOD_LIGHT_UV
] = "uv",
122 [IIO_MOD_LIGHT_UVA
] = "uva",
123 [IIO_MOD_LIGHT_UVB
] = "uvb",
124 [IIO_MOD_LIGHT_DUV
] = "duv",
125 [IIO_MOD_QUATERNION
] = "quaternion",
126 [IIO_MOD_TEMP_AMBIENT
] = "ambient",
127 [IIO_MOD_TEMP_OBJECT
] = "object",
128 [IIO_MOD_NORTH_MAGN
] = "from_north_magnetic",
129 [IIO_MOD_NORTH_TRUE
] = "from_north_true",
130 [IIO_MOD_NORTH_MAGN_TILT_COMP
] = "from_north_magnetic_tilt_comp",
131 [IIO_MOD_NORTH_TRUE_TILT_COMP
] = "from_north_true_tilt_comp",
132 [IIO_MOD_RUNNING
] = "running",
133 [IIO_MOD_JOGGING
] = "jogging",
134 [IIO_MOD_WALKING
] = "walking",
135 [IIO_MOD_STILL
] = "still",
136 [IIO_MOD_ROOT_SUM_SQUARED_X_Y_Z
] = "sqrt(x^2+y^2+z^2)",
139 [IIO_MOD_CO2
] = "co2",
140 [IIO_MOD_VOC
] = "voc",
141 [IIO_MOD_PM1
] = "pm1",
142 [IIO_MOD_PM2P5
] = "pm2p5",
143 [IIO_MOD_PM4
] = "pm4",
144 [IIO_MOD_PM10
] = "pm10",
145 [IIO_MOD_ETHANOL
] = "ethanol",
148 [IIO_MOD_LINEAR_X
] = "linear_x",
149 [IIO_MOD_LINEAR_Y
] = "linear_y",
150 [IIO_MOD_LINEAR_Z
] = "linear_z",
151 [IIO_MOD_PITCH
] = "pitch",
152 [IIO_MOD_YAW
] = "yaw",
153 [IIO_MOD_ROLL
] = "roll",
156 /* relies on pairs of these shared then separate */
157 static const char * const iio_chan_info_postfix
[] = {
158 [IIO_CHAN_INFO_RAW
] = "raw",
159 [IIO_CHAN_INFO_PROCESSED
] = "input",
160 [IIO_CHAN_INFO_SCALE
] = "scale",
161 [IIO_CHAN_INFO_OFFSET
] = "offset",
162 [IIO_CHAN_INFO_CALIBSCALE
] = "calibscale",
163 [IIO_CHAN_INFO_CALIBBIAS
] = "calibbias",
164 [IIO_CHAN_INFO_PEAK
] = "peak_raw",
165 [IIO_CHAN_INFO_PEAK_SCALE
] = "peak_scale",
166 [IIO_CHAN_INFO_QUADRATURE_CORRECTION_RAW
] = "quadrature_correction_raw",
167 [IIO_CHAN_INFO_AVERAGE_RAW
] = "mean_raw",
168 [IIO_CHAN_INFO_LOW_PASS_FILTER_3DB_FREQUENCY
]
169 = "filter_low_pass_3db_frequency",
170 [IIO_CHAN_INFO_HIGH_PASS_FILTER_3DB_FREQUENCY
]
171 = "filter_high_pass_3db_frequency",
172 [IIO_CHAN_INFO_SAMP_FREQ
] = "sampling_frequency",
173 [IIO_CHAN_INFO_FREQUENCY
] = "frequency",
174 [IIO_CHAN_INFO_PHASE
] = "phase",
175 [IIO_CHAN_INFO_HARDWAREGAIN
] = "hardwaregain",
176 [IIO_CHAN_INFO_HYSTERESIS
] = "hysteresis",
177 [IIO_CHAN_INFO_HYSTERESIS_RELATIVE
] = "hysteresis_relative",
178 [IIO_CHAN_INFO_INT_TIME
] = "integration_time",
179 [IIO_CHAN_INFO_ENABLE
] = "en",
180 [IIO_CHAN_INFO_CALIBHEIGHT
] = "calibheight",
181 [IIO_CHAN_INFO_CALIBWEIGHT
] = "calibweight",
182 [IIO_CHAN_INFO_DEBOUNCE_COUNT
] = "debounce_count",
183 [IIO_CHAN_INFO_DEBOUNCE_TIME
] = "debounce_time",
184 [IIO_CHAN_INFO_CALIBEMISSIVITY
] = "calibemissivity",
185 [IIO_CHAN_INFO_OVERSAMPLING_RATIO
] = "oversampling_ratio",
186 [IIO_CHAN_INFO_THERMOCOUPLE_TYPE
] = "thermocouple_type",
187 [IIO_CHAN_INFO_CALIBAMBIENT
] = "calibambient",
188 [IIO_CHAN_INFO_ZEROPOINT
] = "zeropoint",
189 [IIO_CHAN_INFO_TROUGH
] = "trough_raw",
192 * iio_device_id() - query the unique ID for the device
193 * @indio_dev: Device structure whose ID is being queried
195 * The IIO device ID is a unique index used for example for the naming
196 * of the character device /dev/iio\:device[ID].
198 * Returns: Unique ID for the device.
200 int iio_device_id(struct iio_dev
*indio_dev
)
202 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
204 return iio_dev_opaque
->id
;
206 EXPORT_SYMBOL_GPL(iio_device_id
);
209 * iio_buffer_enabled() - helper function to test if the buffer is enabled
210 * @indio_dev: IIO device structure for device
212 * Returns: True, if the buffer is enabled.
214 bool iio_buffer_enabled(struct iio_dev
*indio_dev
)
216 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
218 return iio_dev_opaque
->currentmode
& INDIO_ALL_BUFFER_MODES
;
220 EXPORT_SYMBOL_GPL(iio_buffer_enabled
);
222 #if defined(CONFIG_DEBUG_FS)
224 * There's also a CONFIG_DEBUG_FS guard in include/linux/iio/iio.h for
225 * iio_get_debugfs_dentry() to make it inline if CONFIG_DEBUG_FS is undefined
227 struct dentry
*iio_get_debugfs_dentry(struct iio_dev
*indio_dev
)
229 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
231 return iio_dev_opaque
->debugfs_dentry
;
233 EXPORT_SYMBOL_GPL(iio_get_debugfs_dentry
);
237 * iio_find_channel_from_si() - get channel from its scan index
239 * @si: scan index to match
242 * Constant pointer to iio_chan_spec, if scan index matches, NULL on failure.
244 const struct iio_chan_spec
245 *iio_find_channel_from_si(struct iio_dev
*indio_dev
, int si
)
249 for (i
= 0; i
< indio_dev
->num_channels
; i
++)
250 if (indio_dev
->channels
[i
].scan_index
== si
)
251 return &indio_dev
->channels
[i
];
255 /* This turns up an awful lot */
256 ssize_t
iio_read_const_attr(struct device
*dev
,
257 struct device_attribute
*attr
,
260 return sysfs_emit(buf
, "%s\n", to_iio_const_attr(attr
)->string
);
262 EXPORT_SYMBOL(iio_read_const_attr
);
265 * iio_device_set_clock() - Set current timestamping clock for the device
266 * @indio_dev: IIO device structure containing the device
267 * @clock_id: timestamping clock POSIX identifier to set.
269 * Returns: 0 on success, or a negative error code.
271 int iio_device_set_clock(struct iio_dev
*indio_dev
, clockid_t clock_id
)
274 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
275 const struct iio_event_interface
*ev_int
= iio_dev_opaque
->event_interface
;
277 ret
= mutex_lock_interruptible(&iio_dev_opaque
->mlock
);
280 if ((ev_int
&& iio_event_enabled(ev_int
)) ||
281 iio_buffer_enabled(indio_dev
)) {
282 mutex_unlock(&iio_dev_opaque
->mlock
);
285 iio_dev_opaque
->clock_id
= clock_id
;
286 mutex_unlock(&iio_dev_opaque
->mlock
);
290 EXPORT_SYMBOL(iio_device_set_clock
);
293 * iio_device_get_clock() - Retrieve current timestamping clock for the device
294 * @indio_dev: IIO device structure containing the device
296 * Returns: Clock ID of the current timestamping clock for the device.
298 clockid_t
iio_device_get_clock(const struct iio_dev
*indio_dev
)
300 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
302 return iio_dev_opaque
->clock_id
;
304 EXPORT_SYMBOL(iio_device_get_clock
);
307 * iio_get_time_ns() - utility function to get a time stamp for events etc
310 * Returns: Timestamp of the event in nanoseconds.
312 s64
iio_get_time_ns(const struct iio_dev
*indio_dev
)
314 struct timespec64 tp
;
316 switch (iio_device_get_clock(indio_dev
)) {
318 return ktime_get_real_ns();
319 case CLOCK_MONOTONIC
:
320 return ktime_get_ns();
321 case CLOCK_MONOTONIC_RAW
:
322 return ktime_get_raw_ns();
323 case CLOCK_REALTIME_COARSE
:
324 return ktime_to_ns(ktime_get_coarse_real());
325 case CLOCK_MONOTONIC_COARSE
:
326 ktime_get_coarse_ts64(&tp
);
327 return timespec64_to_ns(&tp
);
329 return ktime_get_boottime_ns();
331 return ktime_get_clocktai_ns();
336 EXPORT_SYMBOL(iio_get_time_ns
);
338 static int __init
iio_init(void)
342 /* Register sysfs bus */
343 ret
= bus_register(&iio_bus_type
);
345 pr_err("could not register bus type\n");
349 ret
= alloc_chrdev_region(&iio_devt
, 0, IIO_DEV_MAX
, "iio");
351 pr_err("failed to allocate char dev region\n");
352 goto error_unregister_bus_type
;
355 iio_debugfs_dentry
= debugfs_create_dir("iio", NULL
);
359 error_unregister_bus_type
:
360 bus_unregister(&iio_bus_type
);
365 static void __exit
iio_exit(void)
368 unregister_chrdev_region(iio_devt
, IIO_DEV_MAX
);
369 bus_unregister(&iio_bus_type
);
370 debugfs_remove(iio_debugfs_dentry
);
373 #if defined(CONFIG_DEBUG_FS)
374 static ssize_t
iio_debugfs_read_reg(struct file
*file
, char __user
*userbuf
,
375 size_t count
, loff_t
*ppos
)
377 struct iio_dev
*indio_dev
= file
->private_data
;
378 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
379 unsigned int val
= 0;
383 return simple_read_from_buffer(userbuf
, count
, ppos
,
384 iio_dev_opaque
->read_buf
,
385 iio_dev_opaque
->read_buf_len
);
387 ret
= indio_dev
->info
->debugfs_reg_access(indio_dev
,
388 iio_dev_opaque
->cached_reg_addr
,
391 dev_err(indio_dev
->dev
.parent
, "%s: read failed\n", __func__
);
395 iio_dev_opaque
->read_buf_len
= snprintf(iio_dev_opaque
->read_buf
,
396 sizeof(iio_dev_opaque
->read_buf
),
399 return simple_read_from_buffer(userbuf
, count
, ppos
,
400 iio_dev_opaque
->read_buf
,
401 iio_dev_opaque
->read_buf_len
);
404 static ssize_t
iio_debugfs_write_reg(struct file
*file
,
405 const char __user
*userbuf
, size_t count
, loff_t
*ppos
)
407 struct iio_dev
*indio_dev
= file
->private_data
;
408 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
409 unsigned int reg
, val
;
413 count
= min(count
, sizeof(buf
) - 1);
414 if (copy_from_user(buf
, userbuf
, count
))
419 ret
= sscanf(buf
, "%i %i", ®
, &val
);
423 iio_dev_opaque
->cached_reg_addr
= reg
;
426 iio_dev_opaque
->cached_reg_addr
= reg
;
427 ret
= indio_dev
->info
->debugfs_reg_access(indio_dev
, reg
,
430 dev_err(indio_dev
->dev
.parent
, "%s: write failed\n",
442 static const struct file_operations iio_debugfs_reg_fops
= {
444 .read
= iio_debugfs_read_reg
,
445 .write
= iio_debugfs_write_reg
,
448 static void iio_device_unregister_debugfs(struct iio_dev
*indio_dev
)
450 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
452 debugfs_remove_recursive(iio_dev_opaque
->debugfs_dentry
);
455 static void iio_device_register_debugfs(struct iio_dev
*indio_dev
)
457 struct iio_dev_opaque
*iio_dev_opaque
;
459 if (indio_dev
->info
->debugfs_reg_access
== NULL
)
462 if (!iio_debugfs_dentry
)
465 iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
467 iio_dev_opaque
->debugfs_dentry
=
468 debugfs_create_dir(dev_name(&indio_dev
->dev
),
471 debugfs_create_file("direct_reg_access", 0644,
472 iio_dev_opaque
->debugfs_dentry
, indio_dev
,
473 &iio_debugfs_reg_fops
);
476 static void iio_device_register_debugfs(struct iio_dev
*indio_dev
)
480 static void iio_device_unregister_debugfs(struct iio_dev
*indio_dev
)
483 #endif /* CONFIG_DEBUG_FS */
485 static ssize_t
iio_read_channel_ext_info(struct device
*dev
,
486 struct device_attribute
*attr
,
489 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
490 struct iio_dev_attr
*this_attr
= to_iio_dev_attr(attr
);
491 const struct iio_chan_spec_ext_info
*ext_info
;
493 ext_info
= &this_attr
->c
->ext_info
[this_attr
->address
];
495 return ext_info
->read(indio_dev
, ext_info
->private, this_attr
->c
, buf
);
498 static ssize_t
iio_write_channel_ext_info(struct device
*dev
,
499 struct device_attribute
*attr
,
500 const char *buf
, size_t len
)
502 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
503 struct iio_dev_attr
*this_attr
= to_iio_dev_attr(attr
);
504 const struct iio_chan_spec_ext_info
*ext_info
;
506 ext_info
= &this_attr
->c
->ext_info
[this_attr
->address
];
508 return ext_info
->write(indio_dev
, ext_info
->private,
509 this_attr
->c
, buf
, len
);
512 ssize_t
iio_enum_available_read(struct iio_dev
*indio_dev
,
513 uintptr_t priv
, const struct iio_chan_spec
*chan
, char *buf
)
515 const struct iio_enum
*e
= (const struct iio_enum
*)priv
;
522 for (i
= 0; i
< e
->num_items
; ++i
) {
525 len
+= sysfs_emit_at(buf
, len
, "%s ", e
->items
[i
]);
528 /* replace last space with a newline */
533 EXPORT_SYMBOL_GPL(iio_enum_available_read
);
535 ssize_t
iio_enum_read(struct iio_dev
*indio_dev
,
536 uintptr_t priv
, const struct iio_chan_spec
*chan
, char *buf
)
538 const struct iio_enum
*e
= (const struct iio_enum
*)priv
;
544 i
= e
->get(indio_dev
, chan
);
547 if (i
>= e
->num_items
|| !e
->items
[i
])
550 return sysfs_emit(buf
, "%s\n", e
->items
[i
]);
552 EXPORT_SYMBOL_GPL(iio_enum_read
);
554 ssize_t
iio_enum_write(struct iio_dev
*indio_dev
,
555 uintptr_t priv
, const struct iio_chan_spec
*chan
, const char *buf
,
558 const struct iio_enum
*e
= (const struct iio_enum
*)priv
;
564 ret
= __sysfs_match_string(e
->items
, e
->num_items
, buf
);
568 ret
= e
->set(indio_dev
, chan
, ret
);
569 return ret
? ret
: len
;
571 EXPORT_SYMBOL_GPL(iio_enum_write
);
573 static const struct iio_mount_matrix iio_mount_idmatrix
= {
581 static int iio_setup_mount_idmatrix(const struct device
*dev
,
582 struct iio_mount_matrix
*matrix
)
584 *matrix
= iio_mount_idmatrix
;
585 dev_info(dev
, "mounting matrix not found: using identity...\n");
589 ssize_t
iio_show_mount_matrix(struct iio_dev
*indio_dev
, uintptr_t priv
,
590 const struct iio_chan_spec
*chan
, char *buf
)
592 const struct iio_mount_matrix
*mtx
;
594 mtx
= ((iio_get_mount_matrix_t
*)priv
)(indio_dev
, chan
);
599 mtx
= &iio_mount_idmatrix
;
601 return sysfs_emit(buf
, "%s, %s, %s; %s, %s, %s; %s, %s, %s\n",
602 mtx
->rotation
[0], mtx
->rotation
[1], mtx
->rotation
[2],
603 mtx
->rotation
[3], mtx
->rotation
[4], mtx
->rotation
[5],
604 mtx
->rotation
[6], mtx
->rotation
[7], mtx
->rotation
[8]);
606 EXPORT_SYMBOL_GPL(iio_show_mount_matrix
);
609 * iio_read_mount_matrix() - retrieve iio device mounting matrix from
610 * device "mount-matrix" property
611 * @dev: device the mounting matrix property is assigned to
612 * @matrix: where to store retrieved matrix
614 * If device is assigned no mounting matrix property, a default 3x3 identity
615 * matrix will be filled in.
617 * Returns: 0 if success, or a negative error code on failure.
619 int iio_read_mount_matrix(struct device
*dev
, struct iio_mount_matrix
*matrix
)
621 size_t len
= ARRAY_SIZE(iio_mount_idmatrix
.rotation
);
624 err
= device_property_read_string_array(dev
, "mount-matrix", matrix
->rotation
, len
);
629 /* Invalid number of matrix entries. */
633 /* Invalid matrix declaration format. */
636 /* Matrix was not declared at all: fallback to identity. */
637 return iio_setup_mount_idmatrix(dev
, matrix
);
639 EXPORT_SYMBOL(iio_read_mount_matrix
);
641 static ssize_t
__iio_format_value(char *buf
, size_t offset
, unsigned int type
,
642 int size
, const int *vals
)
646 bool scale_db
= false;
650 return sysfs_emit_at(buf
, offset
, "%d", vals
[0]);
651 case IIO_VAL_INT_PLUS_MICRO_DB
:
654 case IIO_VAL_INT_PLUS_MICRO
:
656 return sysfs_emit_at(buf
, offset
, "-%d.%06u%s",
657 abs(vals
[0]), -vals
[1],
658 scale_db
? " dB" : "");
660 return sysfs_emit_at(buf
, offset
, "%d.%06u%s", vals
[0],
661 vals
[1], scale_db
? " dB" : "");
662 case IIO_VAL_INT_PLUS_NANO
:
664 return sysfs_emit_at(buf
, offset
, "-%d.%09u",
665 abs(vals
[0]), -vals
[1]);
667 return sysfs_emit_at(buf
, offset
, "%d.%09u", vals
[0],
669 case IIO_VAL_FRACTIONAL
:
670 tmp2
= div_s64((s64
)vals
[0] * 1000000000LL, vals
[1]);
671 tmp0
= (int)div_s64_rem(tmp2
, 1000000000, &tmp1
);
672 if ((tmp2
< 0) && (tmp0
== 0))
673 return sysfs_emit_at(buf
, offset
, "-0.%09u", abs(tmp1
));
675 return sysfs_emit_at(buf
, offset
, "%d.%09u", tmp0
,
677 case IIO_VAL_FRACTIONAL_LOG2
:
678 tmp2
= shift_right((s64
)vals
[0] * 1000000000LL, vals
[1]);
679 tmp0
= (int)div_s64_rem(tmp2
, 1000000000LL, &tmp1
);
680 if (tmp0
== 0 && tmp2
< 0)
681 return sysfs_emit_at(buf
, offset
, "-0.%09u", abs(tmp1
));
683 return sysfs_emit_at(buf
, offset
, "%d.%09u", tmp0
,
685 case IIO_VAL_INT_MULTIPLE
:
690 for (i
= 0; i
< size
; ++i
)
691 l
+= sysfs_emit_at(buf
, offset
+ l
, "%d ", vals
[i
]);
695 return sysfs_emit_at(buf
, offset
, "%c", (char)vals
[0]);
697 tmp2
= (s64
)((((u64
)vals
[1]) << 32) | (u32
)vals
[0]);
698 return sysfs_emit_at(buf
, offset
, "%lld", tmp2
);
705 * iio_format_value() - Formats a IIO value into its string representation
706 * @buf: The buffer to which the formatted value gets written
707 * which is assumed to be big enough (i.e. PAGE_SIZE).
708 * @type: One of the IIO_VAL_* constants. This decides how the val
709 * and val2 parameters are formatted.
710 * @size: Number of IIO value entries contained in vals
711 * @vals: Pointer to the values, exact meaning depends on the
715 * 0 by default, a negative number on failure or the total number of characters
716 * written for a type that belongs to the IIO_VAL_* constant.
718 ssize_t
iio_format_value(char *buf
, unsigned int type
, int size
, int *vals
)
722 len
= __iio_format_value(buf
, 0, type
, size
, vals
);
723 if (len
>= PAGE_SIZE
- 1)
726 return len
+ sysfs_emit_at(buf
, len
, "\n");
728 EXPORT_SYMBOL_GPL(iio_format_value
);
730 ssize_t
do_iio_read_channel_label(struct iio_dev
*indio_dev
,
731 const struct iio_chan_spec
*c
,
734 if (indio_dev
->info
->read_label
)
735 return indio_dev
->info
->read_label(indio_dev
, c
, buf
);
738 return sysfs_emit(buf
, "%s\n", c
->extend_name
);
743 static ssize_t
iio_read_channel_label(struct device
*dev
,
744 struct device_attribute
*attr
,
747 return do_iio_read_channel_label(dev_to_iio_dev(dev
),
748 to_iio_dev_attr(attr
)->c
, buf
);
751 static ssize_t
iio_read_channel_info(struct device
*dev
,
752 struct device_attribute
*attr
,
755 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
756 struct iio_dev_attr
*this_attr
= to_iio_dev_attr(attr
);
757 int vals
[INDIO_MAX_RAW_ELEMENTS
];
761 if (indio_dev
->info
->read_raw_multi
)
762 ret
= indio_dev
->info
->read_raw_multi(indio_dev
, this_attr
->c
,
763 INDIO_MAX_RAW_ELEMENTS
,
766 else if (indio_dev
->info
->read_raw
)
767 ret
= indio_dev
->info
->read_raw(indio_dev
, this_attr
->c
,
768 &vals
[0], &vals
[1], this_attr
->address
);
775 return iio_format_value(buf
, ret
, val_len
, vals
);
778 static ssize_t
iio_format_list(char *buf
, const int *vals
, int type
, int length
,
779 const char *prefix
, const char *suffix
)
794 len
= sysfs_emit(buf
, prefix
);
796 for (i
= 0; i
<= length
- stride
; i
+= stride
) {
798 len
+= sysfs_emit_at(buf
, len
, " ");
799 if (len
>= PAGE_SIZE
)
803 len
+= __iio_format_value(buf
, len
, type
, stride
, &vals
[i
]);
804 if (len
>= PAGE_SIZE
)
808 len
+= sysfs_emit_at(buf
, len
, "%s\n", suffix
);
813 static ssize_t
iio_format_avail_list(char *buf
, const int *vals
,
814 int type
, int length
)
817 return iio_format_list(buf
, vals
, type
, length
, "", "");
820 static ssize_t
iio_format_avail_range(char *buf
, const int *vals
, int type
)
825 * length refers to the array size , not the number of elements.
826 * The purpose is to print the range [min , step ,max] so length should
827 * be 3 in case of int, and 6 for other types.
838 return iio_format_list(buf
, vals
, type
, length
, "[", "]");
841 static ssize_t
iio_read_channel_info_avail(struct device
*dev
,
842 struct device_attribute
*attr
,
845 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
846 struct iio_dev_attr
*this_attr
= to_iio_dev_attr(attr
);
852 if (!indio_dev
->info
->read_avail
)
855 ret
= indio_dev
->info
->read_avail(indio_dev
, this_attr
->c
,
856 &vals
, &type
, &length
,
863 return iio_format_avail_list(buf
, vals
, type
, length
);
864 case IIO_AVAIL_RANGE
:
865 return iio_format_avail_range(buf
, vals
, type
);
872 * __iio_str_to_fixpoint() - Parse a fixed-point number from a string
873 * @str: The string to parse
874 * @fract_mult: Multiplier for the first decimal place, should be a power of 10
875 * @integer: The integer part of the number
876 * @fract: The fractional part of the number
877 * @scale_db: True if this should parse as dB
880 * 0 on success, or a negative error code if the string could not be parsed.
882 static int __iio_str_to_fixpoint(const char *str
, int fract_mult
,
883 int *integer
, int *fract
, bool scale_db
)
886 bool integer_part
= true, negative
= false;
888 if (fract_mult
== 0) {
891 return kstrtoint(str
, 0, integer
);
897 } else if (str
[0] == '+') {
902 if ('0' <= *str
&& *str
<= '9') {
904 i
= i
* 10 + *str
- '0';
906 f
+= fract_mult
* (*str
- '0');
909 } else if (*str
== '\n') {
910 if (*(str
+ 1) == '\0')
913 } else if (!strncmp(str
, " dB", sizeof(" dB") - 1) && scale_db
) {
914 /* Ignore the dB suffix */
915 str
+= sizeof(" dB") - 1;
917 } else if (!strncmp(str
, "dB", sizeof("dB") - 1) && scale_db
) {
918 /* Ignore the dB suffix */
919 str
+= sizeof("dB") - 1;
921 } else if (*str
== '.' && integer_part
) {
922 integer_part
= false;
943 * iio_str_to_fixpoint() - Parse a fixed-point number from a string
944 * @str: The string to parse
945 * @fract_mult: Multiplier for the first decimal place, should be a power of 10
946 * @integer: The integer part of the number
947 * @fract: The fractional part of the number
950 * 0 on success, or a negative error code if the string could not be parsed.
952 int iio_str_to_fixpoint(const char *str
, int fract_mult
,
953 int *integer
, int *fract
)
955 return __iio_str_to_fixpoint(str
, fract_mult
, integer
, fract
, false);
957 EXPORT_SYMBOL_GPL(iio_str_to_fixpoint
);
959 static ssize_t
iio_write_channel_info(struct device
*dev
,
960 struct device_attribute
*attr
,
964 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
965 struct iio_dev_attr
*this_attr
= to_iio_dev_attr(attr
);
966 int ret
, fract_mult
= 100000;
967 int integer
, fract
= 0;
968 bool is_char
= false;
969 bool scale_db
= false;
971 /* Assumes decimal - precision based on number of digits */
972 if (!indio_dev
->info
->write_raw
)
975 if (indio_dev
->info
->write_raw_get_fmt
)
976 switch (indio_dev
->info
->write_raw_get_fmt(indio_dev
,
977 this_attr
->c
, this_attr
->address
)) {
981 case IIO_VAL_INT_PLUS_MICRO_DB
:
984 case IIO_VAL_INT_PLUS_MICRO
:
987 case IIO_VAL_INT_PLUS_NANO
:
988 fract_mult
= 100000000;
1000 if (sscanf(buf
, "%c", &ch
) != 1)
1004 ret
= __iio_str_to_fixpoint(buf
, fract_mult
, &integer
, &fract
,
1010 ret
= indio_dev
->info
->write_raw(indio_dev
, this_attr
->c
,
1011 integer
, fract
, this_attr
->address
);
1019 int __iio_device_attr_init(struct device_attribute
*dev_attr
,
1020 const char *postfix
,
1021 struct iio_chan_spec
const *chan
,
1022 ssize_t (*readfunc
)(struct device
*dev
,
1023 struct device_attribute
*attr
,
1025 ssize_t (*writefunc
)(struct device
*dev
,
1026 struct device_attribute
*attr
,
1029 enum iio_shared_by shared_by
)
1035 sysfs_attr_init(&dev_attr
->attr
);
1037 /* Build up postfix of <extend_name>_<modifier>_postfix */
1038 if (chan
->modified
&& (shared_by
== IIO_SEPARATE
)) {
1039 if (chan
->extend_name
)
1040 full_postfix
= kasprintf(GFP_KERNEL
, "%s_%s_%s",
1041 iio_modifier_names
[chan
->channel2
],
1045 full_postfix
= kasprintf(GFP_KERNEL
, "%s_%s",
1046 iio_modifier_names
[chan
->channel2
],
1049 if (chan
->extend_name
== NULL
|| shared_by
!= IIO_SEPARATE
)
1050 full_postfix
= kstrdup(postfix
, GFP_KERNEL
);
1052 full_postfix
= kasprintf(GFP_KERNEL
,
1057 if (full_postfix
== NULL
)
1060 if (chan
->differential
) { /* Differential can not have modifier */
1061 switch (shared_by
) {
1062 case IIO_SHARED_BY_ALL
:
1063 name
= kasprintf(GFP_KERNEL
, "%s", full_postfix
);
1065 case IIO_SHARED_BY_DIR
:
1066 name
= kasprintf(GFP_KERNEL
, "%s_%s",
1067 iio_direction
[chan
->output
],
1070 case IIO_SHARED_BY_TYPE
:
1071 name
= kasprintf(GFP_KERNEL
, "%s_%s-%s_%s",
1072 iio_direction
[chan
->output
],
1073 iio_chan_type_name_spec
[chan
->type
],
1074 iio_chan_type_name_spec
[chan
->type
],
1078 if (!chan
->indexed
) {
1079 WARN(1, "Differential channels must be indexed\n");
1081 goto error_free_full_postfix
;
1083 name
= kasprintf(GFP_KERNEL
,
1085 iio_direction
[chan
->output
],
1086 iio_chan_type_name_spec
[chan
->type
],
1088 iio_chan_type_name_spec
[chan
->type
],
1093 } else { /* Single ended */
1094 switch (shared_by
) {
1095 case IIO_SHARED_BY_ALL
:
1096 name
= kasprintf(GFP_KERNEL
, "%s", full_postfix
);
1098 case IIO_SHARED_BY_DIR
:
1099 name
= kasprintf(GFP_KERNEL
, "%s_%s",
1100 iio_direction
[chan
->output
],
1103 case IIO_SHARED_BY_TYPE
:
1104 name
= kasprintf(GFP_KERNEL
, "%s_%s_%s",
1105 iio_direction
[chan
->output
],
1106 iio_chan_type_name_spec
[chan
->type
],
1112 name
= kasprintf(GFP_KERNEL
, "%s_%s%d_%s",
1113 iio_direction
[chan
->output
],
1114 iio_chan_type_name_spec
[chan
->type
],
1118 name
= kasprintf(GFP_KERNEL
, "%s_%s_%s",
1119 iio_direction
[chan
->output
],
1120 iio_chan_type_name_spec
[chan
->type
],
1127 goto error_free_full_postfix
;
1129 dev_attr
->attr
.name
= name
;
1132 dev_attr
->attr
.mode
|= 0444;
1133 dev_attr
->show
= readfunc
;
1137 dev_attr
->attr
.mode
|= 0200;
1138 dev_attr
->store
= writefunc
;
1141 error_free_full_postfix
:
1142 kfree(full_postfix
);
1147 static void __iio_device_attr_deinit(struct device_attribute
*dev_attr
)
1149 kfree(dev_attr
->attr
.name
);
1152 int __iio_add_chan_devattr(const char *postfix
,
1153 struct iio_chan_spec
const *chan
,
1154 ssize_t (*readfunc
)(struct device
*dev
,
1155 struct device_attribute
*attr
,
1157 ssize_t (*writefunc
)(struct device
*dev
,
1158 struct device_attribute
*attr
,
1162 enum iio_shared_by shared_by
,
1164 struct iio_buffer
*buffer
,
1165 struct list_head
*attr_list
)
1168 struct iio_dev_attr
*iio_attr
, *t
;
1170 iio_attr
= kzalloc(sizeof(*iio_attr
), GFP_KERNEL
);
1171 if (iio_attr
== NULL
)
1173 ret
= __iio_device_attr_init(&iio_attr
->dev_attr
,
1175 readfunc
, writefunc
, shared_by
);
1177 goto error_iio_dev_attr_free
;
1179 iio_attr
->address
= mask
;
1180 iio_attr
->buffer
= buffer
;
1181 list_for_each_entry(t
, attr_list
, l
)
1182 if (strcmp(t
->dev_attr
.attr
.name
,
1183 iio_attr
->dev_attr
.attr
.name
) == 0) {
1184 if (shared_by
== IIO_SEPARATE
)
1185 dev_err(dev
, "tried to double register : %s\n",
1186 t
->dev_attr
.attr
.name
);
1188 goto error_device_attr_deinit
;
1190 list_add(&iio_attr
->l
, attr_list
);
1194 error_device_attr_deinit
:
1195 __iio_device_attr_deinit(&iio_attr
->dev_attr
);
1196 error_iio_dev_attr_free
:
1201 static int iio_device_add_channel_label(struct iio_dev
*indio_dev
,
1202 struct iio_chan_spec
const *chan
)
1204 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1207 if (!indio_dev
->info
->read_label
&& !chan
->extend_name
)
1210 ret
= __iio_add_chan_devattr("label",
1212 &iio_read_channel_label
,
1218 &iio_dev_opaque
->channel_attr_list
);
1225 static int iio_device_add_info_mask_type(struct iio_dev
*indio_dev
,
1226 struct iio_chan_spec
const *chan
,
1227 enum iio_shared_by shared_by
,
1228 const long *infomask
)
1230 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1231 int i
, ret
, attrcount
= 0;
1233 for_each_set_bit(i
, infomask
, sizeof(*infomask
)*8) {
1234 if (i
>= ARRAY_SIZE(iio_chan_info_postfix
))
1236 ret
= __iio_add_chan_devattr(iio_chan_info_postfix
[i
],
1238 &iio_read_channel_info
,
1239 &iio_write_channel_info
,
1244 &iio_dev_opaque
->channel_attr_list
);
1245 if ((ret
== -EBUSY
) && (shared_by
!= IIO_SEPARATE
))
1255 static int iio_device_add_info_mask_type_avail(struct iio_dev
*indio_dev
,
1256 struct iio_chan_spec
const *chan
,
1257 enum iio_shared_by shared_by
,
1258 const long *infomask
)
1260 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1261 int i
, ret
, attrcount
= 0;
1262 char *avail_postfix
;
1264 for_each_set_bit(i
, infomask
, sizeof(*infomask
) * 8) {
1265 if (i
>= ARRAY_SIZE(iio_chan_info_postfix
))
1267 avail_postfix
= kasprintf(GFP_KERNEL
,
1269 iio_chan_info_postfix
[i
]);
1273 ret
= __iio_add_chan_devattr(avail_postfix
,
1275 &iio_read_channel_info_avail
,
1281 &iio_dev_opaque
->channel_attr_list
);
1282 kfree(avail_postfix
);
1283 if ((ret
== -EBUSY
) && (shared_by
!= IIO_SEPARATE
))
1293 static int iio_device_add_channel_sysfs(struct iio_dev
*indio_dev
,
1294 struct iio_chan_spec
const *chan
)
1296 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1297 int ret
, attrcount
= 0;
1298 const struct iio_chan_spec_ext_info
*ext_info
;
1300 if (chan
->channel
< 0)
1302 ret
= iio_device_add_info_mask_type(indio_dev
, chan
,
1304 &chan
->info_mask_separate
);
1309 ret
= iio_device_add_info_mask_type_avail(indio_dev
, chan
,
1311 &chan
->info_mask_separate_available
);
1316 ret
= iio_device_add_info_mask_type(indio_dev
, chan
,
1318 &chan
->info_mask_shared_by_type
);
1323 ret
= iio_device_add_info_mask_type_avail(indio_dev
, chan
,
1325 &chan
->info_mask_shared_by_type_available
);
1330 ret
= iio_device_add_info_mask_type(indio_dev
, chan
,
1332 &chan
->info_mask_shared_by_dir
);
1337 ret
= iio_device_add_info_mask_type_avail(indio_dev
, chan
,
1339 &chan
->info_mask_shared_by_dir_available
);
1344 ret
= iio_device_add_info_mask_type(indio_dev
, chan
,
1346 &chan
->info_mask_shared_by_all
);
1351 ret
= iio_device_add_info_mask_type_avail(indio_dev
, chan
,
1353 &chan
->info_mask_shared_by_all_available
);
1358 ret
= iio_device_add_channel_label(indio_dev
, chan
);
1363 if (chan
->ext_info
) {
1366 for (ext_info
= chan
->ext_info
; ext_info
->name
; ext_info
++) {
1367 ret
= __iio_add_chan_devattr(ext_info
->name
,
1370 &iio_read_channel_ext_info
: NULL
,
1372 &iio_write_channel_ext_info
: NULL
,
1377 &iio_dev_opaque
->channel_attr_list
);
1379 if (ret
== -EBUSY
&& ext_info
->shared
)
1393 * iio_free_chan_devattr_list() - Free a list of IIO device attributes
1394 * @attr_list: List of IIO device attributes
1396 * This function frees the memory allocated for each of the IIO device
1397 * attributes in the list.
1399 void iio_free_chan_devattr_list(struct list_head
*attr_list
)
1401 struct iio_dev_attr
*p
, *n
;
1403 list_for_each_entry_safe(p
, n
, attr_list
, l
) {
1404 kfree_const(p
->dev_attr
.attr
.name
);
1410 static ssize_t
name_show(struct device
*dev
, struct device_attribute
*attr
,
1413 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
1415 return sysfs_emit(buf
, "%s\n", indio_dev
->name
);
1418 static DEVICE_ATTR_RO(name
);
1420 static ssize_t
label_show(struct device
*dev
, struct device_attribute
*attr
,
1423 struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
1425 return sysfs_emit(buf
, "%s\n", indio_dev
->label
);
1428 static DEVICE_ATTR_RO(label
);
1430 static const char * const clock_names
[] = {
1431 [CLOCK_REALTIME
] = "realtime",
1432 [CLOCK_MONOTONIC
] = "monotonic",
1433 [CLOCK_PROCESS_CPUTIME_ID
] = "process_cputime_id",
1434 [CLOCK_THREAD_CPUTIME_ID
] = "thread_cputime_id",
1435 [CLOCK_MONOTONIC_RAW
] = "monotonic_raw",
1436 [CLOCK_REALTIME_COARSE
] = "realtime_coarse",
1437 [CLOCK_MONOTONIC_COARSE
] = "monotonic_coarse",
1438 [CLOCK_BOOTTIME
] = "boottime",
1439 [CLOCK_REALTIME_ALARM
] = "realtime_alarm",
1440 [CLOCK_BOOTTIME_ALARM
] = "boottime_alarm",
1441 [CLOCK_SGI_CYCLE
] = "sgi_cycle",
1442 [CLOCK_TAI
] = "tai",
1445 static ssize_t
current_timestamp_clock_show(struct device
*dev
,
1446 struct device_attribute
*attr
,
1449 const struct iio_dev
*indio_dev
= dev_to_iio_dev(dev
);
1450 const clockid_t clk
= iio_device_get_clock(indio_dev
);
1453 case CLOCK_REALTIME
:
1454 case CLOCK_MONOTONIC
:
1455 case CLOCK_MONOTONIC_RAW
:
1456 case CLOCK_REALTIME_COARSE
:
1457 case CLOCK_MONOTONIC_COARSE
:
1458 case CLOCK_BOOTTIME
:
1465 return sysfs_emit(buf
, "%s\n", clock_names
[clk
]);
1468 static ssize_t
current_timestamp_clock_store(struct device
*dev
,
1469 struct device_attribute
*attr
,
1470 const char *buf
, size_t len
)
1475 ret
= sysfs_match_string(clock_names
, buf
);
1481 case CLOCK_REALTIME
:
1482 case CLOCK_MONOTONIC
:
1483 case CLOCK_MONOTONIC_RAW
:
1484 case CLOCK_REALTIME_COARSE
:
1485 case CLOCK_MONOTONIC_COARSE
:
1486 case CLOCK_BOOTTIME
:
1493 ret
= iio_device_set_clock(dev_to_iio_dev(dev
), clk
);
1500 int iio_device_register_sysfs_group(struct iio_dev
*indio_dev
,
1501 const struct attribute_group
*group
)
1503 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1504 const struct attribute_group
**new, **old
= iio_dev_opaque
->groups
;
1505 unsigned int cnt
= iio_dev_opaque
->groupcounter
;
1507 new = krealloc_array(old
, cnt
+ 2, sizeof(*new), GFP_KERNEL
);
1511 new[iio_dev_opaque
->groupcounter
++] = group
;
1512 new[iio_dev_opaque
->groupcounter
] = NULL
;
1514 iio_dev_opaque
->groups
= new;
1519 static DEVICE_ATTR_RW(current_timestamp_clock
);
1521 static int iio_device_register_sysfs(struct iio_dev
*indio_dev
)
1523 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1524 int i
, ret
= 0, attrcount
, attrn
, attrcount_orig
= 0;
1525 struct iio_dev_attr
*p
;
1526 struct attribute
**attr
, *clk
= NULL
;
1528 /* First count elements in any existing group */
1529 if (indio_dev
->info
->attrs
) {
1530 attr
= indio_dev
->info
->attrs
->attrs
;
1531 while (*attr
++ != NULL
)
1534 attrcount
= attrcount_orig
;
1536 * New channel registration method - relies on the fact a group does
1537 * not need to be initialized if its name is NULL.
1539 if (indio_dev
->channels
)
1540 for (i
= 0; i
< indio_dev
->num_channels
; i
++) {
1541 const struct iio_chan_spec
*chan
=
1542 &indio_dev
->channels
[i
];
1544 if (chan
->type
== IIO_TIMESTAMP
)
1545 clk
= &dev_attr_current_timestamp_clock
.attr
;
1547 ret
= iio_device_add_channel_sysfs(indio_dev
, chan
);
1549 goto error_clear_attrs
;
1553 if (iio_dev_opaque
->event_interface
)
1554 clk
= &dev_attr_current_timestamp_clock
.attr
;
1556 if (indio_dev
->name
)
1558 if (indio_dev
->label
)
1563 iio_dev_opaque
->chan_attr_group
.attrs
=
1564 kcalloc(attrcount
+ 1,
1565 sizeof(iio_dev_opaque
->chan_attr_group
.attrs
[0]),
1567 if (iio_dev_opaque
->chan_attr_group
.attrs
== NULL
) {
1569 goto error_clear_attrs
;
1571 /* Copy across original attributes, and point to original binary attributes */
1572 if (indio_dev
->info
->attrs
) {
1573 memcpy(iio_dev_opaque
->chan_attr_group
.attrs
,
1574 indio_dev
->info
->attrs
->attrs
,
1575 sizeof(iio_dev_opaque
->chan_attr_group
.attrs
[0])
1577 iio_dev_opaque
->chan_attr_group
.is_visible
=
1578 indio_dev
->info
->attrs
->is_visible
;
1579 iio_dev_opaque
->chan_attr_group
.bin_attrs
=
1580 indio_dev
->info
->attrs
->bin_attrs
;
1582 attrn
= attrcount_orig
;
1583 /* Add all elements from the list. */
1584 list_for_each_entry(p
, &iio_dev_opaque
->channel_attr_list
, l
)
1585 iio_dev_opaque
->chan_attr_group
.attrs
[attrn
++] = &p
->dev_attr
.attr
;
1586 if (indio_dev
->name
)
1587 iio_dev_opaque
->chan_attr_group
.attrs
[attrn
++] = &dev_attr_name
.attr
;
1588 if (indio_dev
->label
)
1589 iio_dev_opaque
->chan_attr_group
.attrs
[attrn
++] = &dev_attr_label
.attr
;
1591 iio_dev_opaque
->chan_attr_group
.attrs
[attrn
++] = clk
;
1593 ret
= iio_device_register_sysfs_group(indio_dev
,
1594 &iio_dev_opaque
->chan_attr_group
);
1596 goto error_free_chan_attrs
;
1600 error_free_chan_attrs
:
1601 kfree(iio_dev_opaque
->chan_attr_group
.attrs
);
1602 iio_dev_opaque
->chan_attr_group
.attrs
= NULL
;
1604 iio_free_chan_devattr_list(&iio_dev_opaque
->channel_attr_list
);
1609 static void iio_device_unregister_sysfs(struct iio_dev
*indio_dev
)
1611 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1613 iio_free_chan_devattr_list(&iio_dev_opaque
->channel_attr_list
);
1614 kfree(iio_dev_opaque
->chan_attr_group
.attrs
);
1615 iio_dev_opaque
->chan_attr_group
.attrs
= NULL
;
1616 kfree(iio_dev_opaque
->groups
);
1617 iio_dev_opaque
->groups
= NULL
;
1620 static void iio_dev_release(struct device
*device
)
1622 struct iio_dev
*indio_dev
= dev_to_iio_dev(device
);
1623 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1625 if (indio_dev
->modes
& INDIO_ALL_TRIGGERED_MODES
)
1626 iio_device_unregister_trigger_consumer(indio_dev
);
1627 iio_device_unregister_eventset(indio_dev
);
1628 iio_device_unregister_sysfs(indio_dev
);
1630 iio_device_detach_buffers(indio_dev
);
1632 lockdep_unregister_key(&iio_dev_opaque
->mlock_key
);
1634 ida_free(&iio_ida
, iio_dev_opaque
->id
);
1635 kfree(iio_dev_opaque
);
1638 const struct device_type iio_device_type
= {
1639 .name
= "iio_device",
1640 .release
= iio_dev_release
,
1644 * iio_device_alloc() - allocate an iio_dev from a driver
1645 * @parent: Parent device.
1646 * @sizeof_priv: Space to allocate for private structure.
1649 * Pointer to allocated iio_dev on success, NULL on failure.
1651 struct iio_dev
*iio_device_alloc(struct device
*parent
, int sizeof_priv
)
1653 struct iio_dev_opaque
*iio_dev_opaque
;
1654 struct iio_dev
*indio_dev
;
1658 alloc_size
= ALIGN(sizeof(*iio_dev_opaque
), IIO_DMA_MINALIGN
) + sizeof_priv
;
1660 alloc_size
= sizeof(*iio_dev_opaque
);
1662 iio_dev_opaque
= kzalloc(alloc_size
, GFP_KERNEL
);
1663 if (!iio_dev_opaque
)
1666 indio_dev
= &iio_dev_opaque
->indio_dev
;
1669 ACCESS_PRIVATE(indio_dev
, priv
) = (char *)iio_dev_opaque
+
1670 ALIGN(sizeof(*iio_dev_opaque
), IIO_DMA_MINALIGN
);
1672 indio_dev
->dev
.parent
= parent
;
1673 indio_dev
->dev
.type
= &iio_device_type
;
1674 indio_dev
->dev
.bus
= &iio_bus_type
;
1675 device_initialize(&indio_dev
->dev
);
1676 mutex_init(&iio_dev_opaque
->mlock
);
1677 mutex_init(&iio_dev_opaque
->info_exist_lock
);
1678 INIT_LIST_HEAD(&iio_dev_opaque
->channel_attr_list
);
1680 iio_dev_opaque
->id
= ida_alloc(&iio_ida
, GFP_KERNEL
);
1681 if (iio_dev_opaque
->id
< 0) {
1682 /* cannot use a dev_err as the name isn't available */
1683 pr_err("failed to get device id\n");
1684 kfree(iio_dev_opaque
);
1688 if (dev_set_name(&indio_dev
->dev
, "iio:device%d", iio_dev_opaque
->id
)) {
1689 ida_free(&iio_ida
, iio_dev_opaque
->id
);
1690 kfree(iio_dev_opaque
);
1694 INIT_LIST_HEAD(&iio_dev_opaque
->buffer_list
);
1695 INIT_LIST_HEAD(&iio_dev_opaque
->ioctl_handlers
);
1697 lockdep_register_key(&iio_dev_opaque
->mlock_key
);
1698 lockdep_set_class(&iio_dev_opaque
->mlock
, &iio_dev_opaque
->mlock_key
);
1702 EXPORT_SYMBOL(iio_device_alloc
);
1705 * iio_device_free() - free an iio_dev from a driver
1706 * @dev: the iio_dev associated with the device
1708 void iio_device_free(struct iio_dev
*dev
)
1711 put_device(&dev
->dev
);
1713 EXPORT_SYMBOL(iio_device_free
);
1715 static void devm_iio_device_release(void *iio_dev
)
1717 iio_device_free(iio_dev
);
1721 * devm_iio_device_alloc - Resource-managed iio_device_alloc()
1722 * @parent: Device to allocate iio_dev for, and parent for this IIO device
1723 * @sizeof_priv: Space to allocate for private structure.
1725 * Managed iio_device_alloc. iio_dev allocated with this function is
1726 * automatically freed on driver detach.
1729 * Pointer to allocated iio_dev on success, NULL on failure.
1731 struct iio_dev
*devm_iio_device_alloc(struct device
*parent
, int sizeof_priv
)
1733 struct iio_dev
*iio_dev
;
1736 iio_dev
= iio_device_alloc(parent
, sizeof_priv
);
1740 ret
= devm_add_action_or_reset(parent
, devm_iio_device_release
,
1747 EXPORT_SYMBOL_GPL(devm_iio_device_alloc
);
1750 * iio_chrdev_open() - chrdev file open for buffer access and ioctls
1751 * @inode: Inode structure for identifying the device in the file system
1752 * @filp: File structure for iio device used to keep and later access
1755 * Returns: 0 on success or -EBUSY if the device is already opened
1757 static int iio_chrdev_open(struct inode
*inode
, struct file
*filp
)
1759 struct iio_dev_opaque
*iio_dev_opaque
=
1760 container_of(inode
->i_cdev
, struct iio_dev_opaque
, chrdev
);
1761 struct iio_dev
*indio_dev
= &iio_dev_opaque
->indio_dev
;
1762 struct iio_dev_buffer_pair
*ib
;
1764 if (test_and_set_bit(IIO_BUSY_BIT_POS
, &iio_dev_opaque
->flags
))
1767 iio_device_get(indio_dev
);
1769 ib
= kmalloc(sizeof(*ib
), GFP_KERNEL
);
1771 iio_device_put(indio_dev
);
1772 clear_bit(IIO_BUSY_BIT_POS
, &iio_dev_opaque
->flags
);
1776 ib
->indio_dev
= indio_dev
;
1777 ib
->buffer
= indio_dev
->buffer
;
1779 filp
->private_data
= ib
;
1785 * iio_chrdev_release() - chrdev file close buffer access and ioctls
1786 * @inode: Inode structure pointer for the char device
1787 * @filp: File structure pointer for the char device
1789 * Returns: 0 for successful release.
1791 static int iio_chrdev_release(struct inode
*inode
, struct file
*filp
)
1793 struct iio_dev_buffer_pair
*ib
= filp
->private_data
;
1794 struct iio_dev_opaque
*iio_dev_opaque
=
1795 container_of(inode
->i_cdev
, struct iio_dev_opaque
, chrdev
);
1796 struct iio_dev
*indio_dev
= &iio_dev_opaque
->indio_dev
;
1799 clear_bit(IIO_BUSY_BIT_POS
, &iio_dev_opaque
->flags
);
1800 iio_device_put(indio_dev
);
1805 void iio_device_ioctl_handler_register(struct iio_dev
*indio_dev
,
1806 struct iio_ioctl_handler
*h
)
1808 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1810 list_add_tail(&h
->entry
, &iio_dev_opaque
->ioctl_handlers
);
1813 void iio_device_ioctl_handler_unregister(struct iio_ioctl_handler
*h
)
1815 list_del(&h
->entry
);
1818 static long iio_ioctl(struct file
*filp
, unsigned int cmd
, unsigned long arg
)
1820 struct iio_dev_buffer_pair
*ib
= filp
->private_data
;
1821 struct iio_dev
*indio_dev
= ib
->indio_dev
;
1822 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
1823 struct iio_ioctl_handler
*h
;
1826 guard(mutex
)(&iio_dev_opaque
->info_exist_lock
);
1828 * The NULL check here is required to prevent crashing when a device
1829 * is being removed while userspace would still have open file handles
1830 * to try to access this device.
1832 if (!indio_dev
->info
)
1835 list_for_each_entry(h
, &iio_dev_opaque
->ioctl_handlers
, entry
) {
1836 ret
= h
->ioctl(indio_dev
, filp
, cmd
, arg
);
1837 if (ret
!= IIO_IOCTL_UNHANDLED
)
1844 static const struct file_operations iio_buffer_fileops
= {
1845 .owner
= THIS_MODULE
,
1846 .llseek
= noop_llseek
,
1847 .read
= iio_buffer_read_outer_addr
,
1848 .write
= iio_buffer_write_outer_addr
,
1849 .poll
= iio_buffer_poll_addr
,
1850 .unlocked_ioctl
= iio_ioctl
,
1851 .compat_ioctl
= compat_ptr_ioctl
,
1852 .open
= iio_chrdev_open
,
1853 .release
= iio_chrdev_release
,
1856 static const struct file_operations iio_event_fileops
= {
1857 .owner
= THIS_MODULE
,
1858 .llseek
= noop_llseek
,
1859 .unlocked_ioctl
= iio_ioctl
,
1860 .compat_ioctl
= compat_ptr_ioctl
,
1861 .open
= iio_chrdev_open
,
1862 .release
= iio_chrdev_release
,
1865 static int iio_check_unique_scan_index(struct iio_dev
*indio_dev
)
1868 const struct iio_chan_spec
*channels
= indio_dev
->channels
;
1870 if (!(indio_dev
->modes
& INDIO_ALL_BUFFER_MODES
))
1873 for (i
= 0; i
< indio_dev
->num_channels
- 1; i
++) {
1874 if (channels
[i
].scan_index
< 0)
1876 for (j
= i
+ 1; j
< indio_dev
->num_channels
; j
++)
1877 if (channels
[i
].scan_index
== channels
[j
].scan_index
) {
1878 dev_err(&indio_dev
->dev
,
1879 "Duplicate scan index %d\n",
1880 channels
[i
].scan_index
);
1888 static int iio_check_extended_name(const struct iio_dev
*indio_dev
)
1892 if (!indio_dev
->info
->read_label
)
1895 for (i
= 0; i
< indio_dev
->num_channels
; i
++) {
1896 if (indio_dev
->channels
[i
].extend_name
) {
1897 dev_err(&indio_dev
->dev
,
1898 "Cannot use labels and extend_name at the same time\n");
1906 static const struct iio_buffer_setup_ops noop_ring_setup_ops
;
1908 static void iio_sanity_check_avail_scan_masks(struct iio_dev
*indio_dev
)
1910 unsigned int num_masks
, masklength
, longs_per_mask
;
1911 const unsigned long *av_masks
;
1914 av_masks
= indio_dev
->available_scan_masks
;
1915 masklength
= iio_get_masklength(indio_dev
);
1916 longs_per_mask
= BITS_TO_LONGS(masklength
);
1919 * The code determining how many available_scan_masks is in the array
1920 * will be assuming the end of masks when first long with all bits
1921 * zeroed is encountered. This is incorrect for masks where mask
1922 * consists of more than one long, and where some of the available masks
1923 * has long worth of bits zeroed (but has subsequent bit(s) set). This
1924 * is a safety measure against bug where array of masks is terminated by
1925 * a single zero while mask width is greater than width of a long.
1927 if (longs_per_mask
> 1)
1928 dev_warn(indio_dev
->dev
.parent
,
1929 "multi long available scan masks not fully supported\n");
1931 if (bitmap_empty(av_masks
, masklength
))
1932 dev_warn(indio_dev
->dev
.parent
, "empty scan mask\n");
1934 for (num_masks
= 0; *av_masks
; num_masks
++)
1935 av_masks
+= longs_per_mask
;
1940 av_masks
= indio_dev
->available_scan_masks
;
1943 * Go through all the masks from first to one before the last, and see
1944 * that no mask found later from the available_scan_masks array is a
1945 * subset of mask found earlier. If this happens, then the mask found
1946 * later will never get used because scanning the array is stopped when
1947 * the first suitable mask is found. Drivers should order the array of
1948 * available masks in the order of preference (presumably the least
1949 * costy to access masks first).
1951 for (i
= 0; i
< num_masks
- 1; i
++) {
1952 const unsigned long *mask1
;
1955 mask1
= av_masks
+ i
* longs_per_mask
;
1956 for (j
= i
+ 1; j
< num_masks
; j
++) {
1957 const unsigned long *mask2
;
1959 mask2
= av_masks
+ j
* longs_per_mask
;
1960 if (bitmap_subset(mask2
, mask1
, masklength
))
1961 dev_warn(indio_dev
->dev
.parent
,
1962 "available_scan_mask %d subset of %d. Never used\n",
1969 * iio_active_scan_mask_index - Get index of the active scan mask inside the
1970 * available scan masks array
1971 * @indio_dev: the IIO device containing the active and available scan masks
1973 * Returns: the index or -EINVAL if active_scan_mask is not set
1975 int iio_active_scan_mask_index(struct iio_dev
*indio_dev
)
1978 const unsigned long *av_masks
;
1979 unsigned int masklength
= iio_get_masklength(indio_dev
);
1982 if (!indio_dev
->active_scan_mask
)
1986 * As in iio_scan_mask_match and iio_sanity_check_avail_scan_masks,
1987 * the condition here do not handle multi-long masks correctly.
1988 * It only checks the first long to be zero, and will use such mask
1989 * as a terminator even if there was bits set after the first long.
1991 * This should be fine since the available_scan_mask has already been
1992 * sanity tested using iio_sanity_check_avail_scan_masks.
1994 * See iio_scan_mask_match and iio_sanity_check_avail_scan_masks for
1997 av_masks
= indio_dev
->available_scan_masks
;
1999 if (indio_dev
->active_scan_mask
== av_masks
)
2001 av_masks
+= BITS_TO_LONGS(masklength
);
2005 dev_warn(indio_dev
->dev
.parent
,
2006 "active scan mask is not part of the available scan masks\n");
2009 EXPORT_SYMBOL_GPL(iio_active_scan_mask_index
);
2011 int __iio_device_register(struct iio_dev
*indio_dev
, struct module
*this_mod
)
2013 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
2014 struct fwnode_handle
*fwnode
= NULL
;
2017 if (!indio_dev
->info
)
2020 iio_dev_opaque
->driver_module
= this_mod
;
2022 /* If the calling driver did not initialize firmware node, do it here */
2023 if (dev_fwnode(&indio_dev
->dev
))
2024 fwnode
= dev_fwnode(&indio_dev
->dev
);
2025 /* The default dummy IIO device has no parent */
2026 else if (indio_dev
->dev
.parent
)
2027 fwnode
= dev_fwnode(indio_dev
->dev
.parent
);
2028 device_set_node(&indio_dev
->dev
, fwnode
);
2030 fwnode_property_read_string(fwnode
, "label", &indio_dev
->label
);
2032 ret
= iio_check_unique_scan_index(indio_dev
);
2036 ret
= iio_check_extended_name(indio_dev
);
2040 iio_device_register_debugfs(indio_dev
);
2042 ret
= iio_buffers_alloc_sysfs_and_mask(indio_dev
);
2044 dev_err(indio_dev
->dev
.parent
,
2045 "Failed to create buffer sysfs interfaces\n");
2046 goto error_unreg_debugfs
;
2049 if (indio_dev
->available_scan_masks
)
2050 iio_sanity_check_avail_scan_masks(indio_dev
);
2052 ret
= iio_device_register_sysfs(indio_dev
);
2054 dev_err(indio_dev
->dev
.parent
,
2055 "Failed to register sysfs interfaces\n");
2056 goto error_buffer_free_sysfs
;
2058 ret
= iio_device_register_eventset(indio_dev
);
2060 dev_err(indio_dev
->dev
.parent
,
2061 "Failed to register event set\n");
2062 goto error_free_sysfs
;
2064 if (indio_dev
->modes
& INDIO_ALL_TRIGGERED_MODES
)
2065 iio_device_register_trigger_consumer(indio_dev
);
2067 if ((indio_dev
->modes
& INDIO_ALL_BUFFER_MODES
) &&
2068 indio_dev
->setup_ops
== NULL
)
2069 indio_dev
->setup_ops
= &noop_ring_setup_ops
;
2071 if (iio_dev_opaque
->attached_buffers_cnt
)
2072 cdev_init(&iio_dev_opaque
->chrdev
, &iio_buffer_fileops
);
2073 else if (iio_dev_opaque
->event_interface
)
2074 cdev_init(&iio_dev_opaque
->chrdev
, &iio_event_fileops
);
2076 if (iio_dev_opaque
->attached_buffers_cnt
|| iio_dev_opaque
->event_interface
) {
2077 indio_dev
->dev
.devt
= MKDEV(MAJOR(iio_devt
), iio_dev_opaque
->id
);
2078 iio_dev_opaque
->chrdev
.owner
= this_mod
;
2081 /* assign device groups now; they should be all registered now */
2082 indio_dev
->dev
.groups
= iio_dev_opaque
->groups
;
2084 ret
= cdev_device_add(&iio_dev_opaque
->chrdev
, &indio_dev
->dev
);
2086 goto error_unreg_eventset
;
2090 error_unreg_eventset
:
2091 iio_device_unregister_eventset(indio_dev
);
2093 iio_device_unregister_sysfs(indio_dev
);
2094 error_buffer_free_sysfs
:
2095 iio_buffers_free_sysfs_and_mask(indio_dev
);
2096 error_unreg_debugfs
:
2097 iio_device_unregister_debugfs(indio_dev
);
2100 EXPORT_SYMBOL(__iio_device_register
);
2103 * iio_device_unregister() - unregister a device from the IIO subsystem
2104 * @indio_dev: Device structure representing the device.
2106 void iio_device_unregister(struct iio_dev
*indio_dev
)
2108 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
2110 cdev_device_del(&iio_dev_opaque
->chrdev
, &indio_dev
->dev
);
2112 scoped_guard(mutex
, &iio_dev_opaque
->info_exist_lock
) {
2113 iio_device_unregister_debugfs(indio_dev
);
2115 iio_disable_all_buffers(indio_dev
);
2117 indio_dev
->info
= NULL
;
2119 iio_device_wakeup_eventset(indio_dev
);
2120 iio_buffer_wakeup_poll(indio_dev
);
2123 iio_buffers_free_sysfs_and_mask(indio_dev
);
2125 EXPORT_SYMBOL(iio_device_unregister
);
2127 static void devm_iio_device_unreg(void *indio_dev
)
2129 iio_device_unregister(indio_dev
);
2132 int __devm_iio_device_register(struct device
*dev
, struct iio_dev
*indio_dev
,
2133 struct module
*this_mod
)
2137 ret
= __iio_device_register(indio_dev
, this_mod
);
2141 return devm_add_action_or_reset(dev
, devm_iio_device_unreg
, indio_dev
);
2143 EXPORT_SYMBOL_GPL(__devm_iio_device_register
);
2146 * iio_device_claim_direct_mode - Keep device in direct mode
2147 * @indio_dev: the iio_dev associated with the device
2149 * If the device is in direct mode it is guaranteed to stay
2150 * that way until iio_device_release_direct_mode() is called.
2152 * Use with iio_device_release_direct_mode()
2154 * Returns: 0 on success, -EBUSY on failure.
2156 int iio_device_claim_direct_mode(struct iio_dev
*indio_dev
)
2158 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
2160 mutex_lock(&iio_dev_opaque
->mlock
);
2162 if (iio_buffer_enabled(indio_dev
)) {
2163 mutex_unlock(&iio_dev_opaque
->mlock
);
2168 EXPORT_SYMBOL_GPL(iio_device_claim_direct_mode
);
2171 * iio_device_release_direct_mode - releases claim on direct mode
2172 * @indio_dev: the iio_dev associated with the device
2174 * Release the claim. Device is no longer guaranteed to stay
2177 * Use with iio_device_claim_direct_mode()
2179 void iio_device_release_direct_mode(struct iio_dev
*indio_dev
)
2181 mutex_unlock(&to_iio_dev_opaque(indio_dev
)->mlock
);
2183 EXPORT_SYMBOL_GPL(iio_device_release_direct_mode
);
2186 * iio_device_claim_buffer_mode - Keep device in buffer mode
2187 * @indio_dev: the iio_dev associated with the device
2189 * If the device is in buffer mode it is guaranteed to stay
2190 * that way until iio_device_release_buffer_mode() is called.
2192 * Use with iio_device_release_buffer_mode().
2194 * Returns: 0 on success, -EBUSY on failure.
2196 int iio_device_claim_buffer_mode(struct iio_dev
*indio_dev
)
2198 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
2200 mutex_lock(&iio_dev_opaque
->mlock
);
2202 if (iio_buffer_enabled(indio_dev
))
2205 mutex_unlock(&iio_dev_opaque
->mlock
);
2208 EXPORT_SYMBOL_GPL(iio_device_claim_buffer_mode
);
2211 * iio_device_release_buffer_mode - releases claim on buffer mode
2212 * @indio_dev: the iio_dev associated with the device
2214 * Release the claim. Device is no longer guaranteed to stay
2217 * Use with iio_device_claim_buffer_mode().
2219 void iio_device_release_buffer_mode(struct iio_dev
*indio_dev
)
2221 mutex_unlock(&to_iio_dev_opaque(indio_dev
)->mlock
);
2223 EXPORT_SYMBOL_GPL(iio_device_release_buffer_mode
);
2226 * iio_device_get_current_mode() - helper function providing read-only access to
2227 * the opaque @currentmode variable
2228 * @indio_dev: IIO device structure for device
2230 int iio_device_get_current_mode(struct iio_dev
*indio_dev
)
2232 struct iio_dev_opaque
*iio_dev_opaque
= to_iio_dev_opaque(indio_dev
);
2234 return iio_dev_opaque
->currentmode
;
2236 EXPORT_SYMBOL_GPL(iio_device_get_current_mode
);
2238 subsys_initcall(iio_init
);
2239 module_exit(iio_exit
);
2241 MODULE_AUTHOR("Jonathan Cameron <jic23@kernel.org>");
2242 MODULE_DESCRIPTION("Industrial I/O core");
2243 MODULE_LICENSE("GPL");