.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-only |
---|
1 | 2 | /* The industrial I/O core |
---|
2 | 3 | * |
---|
3 | 4 | * Copyright (c) 2008 Jonathan Cameron |
---|
4 | 5 | * |
---|
5 | | - * This program is free software; you can redistribute it and/or modify it |
---|
6 | | - * under the terms of the GNU General Public License version 2 as published by |
---|
7 | | - * the Free Software Foundation. |
---|
8 | | - * |
---|
9 | 6 | * Handling of buffer allocation / resizing. |
---|
10 | | - * |
---|
11 | 7 | * |
---|
12 | 8 | * Things to look at here. |
---|
13 | 9 | * - Better memory allocation techniques? |
---|
.. | .. |
---|
23 | 19 | #include <linux/sched/signal.h> |
---|
24 | 20 | |
---|
25 | 21 | #include <linux/iio/iio.h> |
---|
| 22 | +#include <linux/iio/iio-opaque.h> |
---|
26 | 23 | #include "iio_core.h" |
---|
| 24 | +#include "iio_core_trigger.h" |
---|
27 | 25 | #include <linux/iio/sysfs.h> |
---|
28 | 26 | #include <linux/iio/buffer.h> |
---|
29 | 27 | #include <linux/iio/buffer_impl.h> |
---|
.. | .. |
---|
91 | 89 | } |
---|
92 | 90 | |
---|
93 | 91 | /** |
---|
94 | | - * iio_buffer_read_first_n_outer() - chrdev read for buffer access |
---|
| 92 | + * iio_buffer_read_outer() - chrdev read for buffer access |
---|
95 | 93 | * @filp: File structure pointer for the char device |
---|
96 | 94 | * @buf: Destination buffer for iio buffer read |
---|
97 | 95 | * @n: First n bytes to read |
---|
.. | .. |
---|
103 | 101 | * Return: negative values corresponding to error codes or ret != 0 |
---|
104 | 102 | * for ending the reading activity |
---|
105 | 103 | **/ |
---|
106 | | -ssize_t iio_buffer_read_first_n_outer(struct file *filp, char __user *buf, |
---|
107 | | - size_t n, loff_t *f_ps) |
---|
| 104 | +ssize_t iio_buffer_read_outer(struct file *filp, char __user *buf, |
---|
| 105 | + size_t n, loff_t *f_ps) |
---|
108 | 106 | { |
---|
109 | 107 | struct iio_dev *indio_dev = filp->private_data; |
---|
110 | 108 | struct iio_buffer *rb = indio_dev->buffer; |
---|
.. | .. |
---|
116 | 114 | if (!indio_dev->info) |
---|
117 | 115 | return -ENODEV; |
---|
118 | 116 | |
---|
119 | | - if (!rb || !rb->access->read_first_n) |
---|
| 117 | + if (!rb || !rb->access->read) |
---|
120 | 118 | return -EINVAL; |
---|
121 | 119 | |
---|
122 | 120 | datum_size = rb->bytes_per_datum; |
---|
.. | .. |
---|
151 | 149 | continue; |
---|
152 | 150 | } |
---|
153 | 151 | |
---|
154 | | - ret = rb->access->read_first_n(rb, n, buf); |
---|
| 152 | + ret = rb->access->read(rb, n, buf); |
---|
155 | 153 | if (ret == 0 && (filp->f_flags & O_NONBLOCK)) |
---|
156 | 154 | ret = -EAGAIN; |
---|
157 | 155 | } while (ret == 0); |
---|
.. | .. |
---|
193 | 191 | */ |
---|
194 | 192 | void iio_buffer_wakeup_poll(struct iio_dev *indio_dev) |
---|
195 | 193 | { |
---|
196 | | - if (!indio_dev->buffer) |
---|
| 194 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
| 195 | + |
---|
| 196 | + if (!buffer) |
---|
197 | 197 | return; |
---|
198 | 198 | |
---|
199 | | - wake_up(&indio_dev->buffer->pollq); |
---|
| 199 | + wake_up(&buffer->pollq); |
---|
200 | 200 | } |
---|
201 | 201 | |
---|
202 | 202 | void iio_buffer_init(struct iio_buffer *buffer) |
---|
.. | .. |
---|
266 | 266 | { |
---|
267 | 267 | int ret; |
---|
268 | 268 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
| 269 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
269 | 270 | |
---|
270 | 271 | /* Ensure ret is 0 or 1. */ |
---|
271 | 272 | ret = !!test_bit(to_iio_dev_attr(attr)->address, |
---|
272 | | - indio_dev->buffer->scan_mask); |
---|
| 273 | + buffer->scan_mask); |
---|
273 | 274 | |
---|
274 | 275 | return sprintf(buf, "%d\n", ret); |
---|
275 | 276 | } |
---|
.. | .. |
---|
320 | 321 | const unsigned long *mask; |
---|
321 | 322 | unsigned long *trialmask; |
---|
322 | 323 | |
---|
323 | | - trialmask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
---|
324 | | - sizeof(*trialmask), GFP_KERNEL); |
---|
| 324 | + trialmask = bitmap_zalloc(indio_dev->masklength, GFP_KERNEL); |
---|
325 | 325 | if (trialmask == NULL) |
---|
326 | 326 | return -ENOMEM; |
---|
327 | 327 | if (!indio_dev->masklength) { |
---|
.. | .. |
---|
343 | 343 | } |
---|
344 | 344 | bitmap_copy(buffer->scan_mask, trialmask, indio_dev->masklength); |
---|
345 | 345 | |
---|
346 | | - kfree(trialmask); |
---|
| 346 | + bitmap_free(trialmask); |
---|
347 | 347 | |
---|
348 | 348 | return 0; |
---|
349 | 349 | |
---|
350 | 350 | err_invalid_mask: |
---|
351 | | - kfree(trialmask); |
---|
| 351 | + bitmap_free(trialmask); |
---|
352 | 352 | return -EINVAL; |
---|
353 | 353 | } |
---|
354 | 354 | |
---|
.. | .. |
---|
386 | 386 | if (ret < 0) |
---|
387 | 387 | return ret; |
---|
388 | 388 | mutex_lock(&indio_dev->mlock); |
---|
389 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
---|
| 389 | + if (iio_buffer_is_active(buffer)) { |
---|
390 | 390 | ret = -EBUSY; |
---|
391 | 391 | goto error_ret; |
---|
392 | 392 | } |
---|
.. | .. |
---|
415 | 415 | char *buf) |
---|
416 | 416 | { |
---|
417 | 417 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
418 | | - return sprintf(buf, "%d\n", indio_dev->buffer->scan_timestamp); |
---|
| 418 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
| 419 | + |
---|
| 420 | + return sprintf(buf, "%d\n", buffer->scan_timestamp); |
---|
419 | 421 | } |
---|
420 | 422 | |
---|
421 | 423 | static ssize_t iio_scan_el_ts_store(struct device *dev, |
---|
.. | .. |
---|
425 | 427 | { |
---|
426 | 428 | int ret; |
---|
427 | 429 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
| 430 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
428 | 431 | bool state; |
---|
429 | 432 | |
---|
430 | 433 | ret = strtobool(buf, &state); |
---|
.. | .. |
---|
432 | 435 | return ret; |
---|
433 | 436 | |
---|
434 | 437 | mutex_lock(&indio_dev->mlock); |
---|
435 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
---|
| 438 | + if (iio_buffer_is_active(buffer)) { |
---|
436 | 439 | ret = -EBUSY; |
---|
437 | 440 | goto error_ret; |
---|
438 | 441 | } |
---|
439 | | - indio_dev->buffer->scan_timestamp = state; |
---|
| 442 | + buffer->scan_timestamp = state; |
---|
440 | 443 | error_ret: |
---|
441 | 444 | mutex_unlock(&indio_dev->mlock); |
---|
442 | 445 | |
---|
.. | .. |
---|
444 | 447 | } |
---|
445 | 448 | |
---|
446 | 449 | static int iio_buffer_add_channel_sysfs(struct iio_dev *indio_dev, |
---|
| 450 | + struct iio_buffer *buffer, |
---|
447 | 451 | const struct iio_chan_spec *chan) |
---|
448 | 452 | { |
---|
449 | 453 | int ret, attrcount = 0; |
---|
450 | | - struct iio_buffer *buffer = indio_dev->buffer; |
---|
451 | 454 | |
---|
452 | 455 | ret = __iio_add_chan_devattr("index", |
---|
453 | 456 | chan, |
---|
.. | .. |
---|
523 | 526 | return len; |
---|
524 | 527 | |
---|
525 | 528 | mutex_lock(&indio_dev->mlock); |
---|
526 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
---|
| 529 | + if (iio_buffer_is_active(buffer)) { |
---|
527 | 530 | ret = -EBUSY; |
---|
528 | 531 | } else { |
---|
529 | 532 | buffer->access->set_length(buffer, val); |
---|
.. | .. |
---|
544 | 547 | char *buf) |
---|
545 | 548 | { |
---|
546 | 549 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
547 | | - return sprintf(buf, "%d\n", iio_buffer_is_active(indio_dev->buffer)); |
---|
| 550 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
| 551 | + |
---|
| 552 | + return sprintf(buf, "%d\n", iio_buffer_is_active(buffer)); |
---|
548 | 553 | } |
---|
549 | 554 | |
---|
550 | 555 | static unsigned int iio_storage_bytes_for_si(struct iio_dev *indio_dev, |
---|
.. | .. |
---|
595 | 600 | static void iio_buffer_activate(struct iio_dev *indio_dev, |
---|
596 | 601 | struct iio_buffer *buffer) |
---|
597 | 602 | { |
---|
| 603 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
| 604 | + |
---|
598 | 605 | iio_buffer_get(buffer); |
---|
599 | | - list_add(&buffer->buffer_list, &indio_dev->buffer_list); |
---|
| 606 | + list_add(&buffer->buffer_list, &iio_dev_opaque->buffer_list); |
---|
600 | 607 | } |
---|
601 | 608 | |
---|
602 | 609 | static void iio_buffer_deactivate(struct iio_buffer *buffer) |
---|
.. | .. |
---|
608 | 615 | |
---|
609 | 616 | static void iio_buffer_deactivate_all(struct iio_dev *indio_dev) |
---|
610 | 617 | { |
---|
| 618 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
611 | 619 | struct iio_buffer *buffer, *_buffer; |
---|
612 | 620 | |
---|
613 | 621 | list_for_each_entry_safe(buffer, _buffer, |
---|
614 | | - &indio_dev->buffer_list, buffer_list) |
---|
| 622 | + &iio_dev_opaque->buffer_list, buffer_list) |
---|
615 | 623 | iio_buffer_deactivate(buffer); |
---|
616 | 624 | } |
---|
617 | 625 | |
---|
.. | .. |
---|
669 | 677 | { |
---|
670 | 678 | /* If the mask is dynamically allocated free it, otherwise do nothing */ |
---|
671 | 679 | if (!indio_dev->available_scan_masks) |
---|
672 | | - kfree(mask); |
---|
| 680 | + bitmap_free(mask); |
---|
673 | 681 | } |
---|
674 | 682 | |
---|
675 | 683 | struct iio_device_config { |
---|
.. | .. |
---|
684 | 692 | struct iio_buffer *insert_buffer, struct iio_buffer *remove_buffer, |
---|
685 | 693 | struct iio_device_config *config) |
---|
686 | 694 | { |
---|
| 695 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
687 | 696 | unsigned long *compound_mask; |
---|
688 | 697 | const unsigned long *scan_mask; |
---|
689 | 698 | bool strict_scanmask = false; |
---|
690 | 699 | struct iio_buffer *buffer; |
---|
691 | 700 | bool scan_timestamp; |
---|
692 | 701 | unsigned int modes; |
---|
| 702 | + |
---|
| 703 | + if (insert_buffer && |
---|
| 704 | + bitmap_empty(insert_buffer->scan_mask, indio_dev->masklength)) { |
---|
| 705 | + dev_dbg(&indio_dev->dev, |
---|
| 706 | + "At least one scan element must be enabled first\n"); |
---|
| 707 | + return -EINVAL; |
---|
| 708 | + } |
---|
693 | 709 | |
---|
694 | 710 | memset(config, 0, sizeof(*config)); |
---|
695 | 711 | config->watermark = ~0; |
---|
.. | .. |
---|
699 | 715 | * to verify. |
---|
700 | 716 | */ |
---|
701 | 717 | if (remove_buffer && !insert_buffer && |
---|
702 | | - list_is_singular(&indio_dev->buffer_list)) |
---|
| 718 | + list_is_singular(&iio_dev_opaque->buffer_list)) |
---|
703 | 719 | return 0; |
---|
704 | 720 | |
---|
705 | 721 | modes = indio_dev->modes; |
---|
706 | 722 | |
---|
707 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
---|
| 723 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
708 | 724 | if (buffer == remove_buffer) |
---|
709 | 725 | continue; |
---|
710 | 726 | modes &= buffer->access->modes; |
---|
.. | .. |
---|
725 | 741 | * Keep things simple for now and only allow a single buffer to |
---|
726 | 742 | * be connected in hardware mode. |
---|
727 | 743 | */ |
---|
728 | | - if (insert_buffer && !list_empty(&indio_dev->buffer_list)) |
---|
| 744 | + if (insert_buffer && !list_empty(&iio_dev_opaque->buffer_list)) |
---|
729 | 745 | return -EINVAL; |
---|
730 | 746 | config->mode = INDIO_BUFFER_HARDWARE; |
---|
731 | 747 | strict_scanmask = true; |
---|
.. | .. |
---|
739 | 755 | } |
---|
740 | 756 | |
---|
741 | 757 | /* What scan mask do we actually have? */ |
---|
742 | | - compound_mask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
---|
743 | | - sizeof(long), GFP_KERNEL); |
---|
| 758 | + compound_mask = bitmap_zalloc(indio_dev->masklength, GFP_KERNEL); |
---|
744 | 759 | if (compound_mask == NULL) |
---|
745 | 760 | return -ENOMEM; |
---|
746 | 761 | |
---|
747 | 762 | scan_timestamp = false; |
---|
748 | 763 | |
---|
749 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
---|
| 764 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
750 | 765 | if (buffer == remove_buffer) |
---|
751 | 766 | continue; |
---|
752 | 767 | bitmap_or(compound_mask, compound_mask, buffer->scan_mask, |
---|
.. | .. |
---|
765 | 780 | indio_dev->masklength, |
---|
766 | 781 | compound_mask, |
---|
767 | 782 | strict_scanmask); |
---|
768 | | - kfree(compound_mask); |
---|
| 783 | + bitmap_free(compound_mask); |
---|
769 | 784 | if (scan_mask == NULL) |
---|
770 | 785 | return -EINVAL; |
---|
771 | 786 | } else { |
---|
.. | .. |
---|
892 | 907 | |
---|
893 | 908 | static int iio_update_demux(struct iio_dev *indio_dev) |
---|
894 | 909 | { |
---|
| 910 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
895 | 911 | struct iio_buffer *buffer; |
---|
896 | 912 | int ret; |
---|
897 | 913 | |
---|
898 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
---|
| 914 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
899 | 915 | ret = iio_buffer_update_demux(indio_dev, buffer); |
---|
900 | 916 | if (ret < 0) |
---|
901 | 917 | goto error_clear_mux_table; |
---|
.. | .. |
---|
903 | 919 | return 0; |
---|
904 | 920 | |
---|
905 | 921 | error_clear_mux_table: |
---|
906 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) |
---|
| 922 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) |
---|
907 | 923 | iio_buffer_demux_free(buffer); |
---|
908 | 924 | |
---|
909 | 925 | return ret; |
---|
.. | .. |
---|
912 | 928 | static int iio_enable_buffers(struct iio_dev *indio_dev, |
---|
913 | 929 | struct iio_device_config *config) |
---|
914 | 930 | { |
---|
| 931 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
915 | 932 | struct iio_buffer *buffer; |
---|
916 | 933 | int ret; |
---|
917 | 934 | |
---|
918 | 935 | indio_dev->active_scan_mask = config->scan_mask; |
---|
919 | 936 | indio_dev->scan_timestamp = config->scan_timestamp; |
---|
920 | 937 | indio_dev->scan_bytes = config->scan_bytes; |
---|
| 938 | + indio_dev->currentmode = config->mode; |
---|
921 | 939 | |
---|
922 | 940 | iio_update_demux(indio_dev); |
---|
923 | 941 | |
---|
.. | .. |
---|
947 | 965 | indio_dev->info->hwfifo_set_watermark(indio_dev, |
---|
948 | 966 | config->watermark); |
---|
949 | 967 | |
---|
950 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
---|
| 968 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
951 | 969 | ret = iio_buffer_enable(buffer, indio_dev); |
---|
952 | 970 | if (ret) |
---|
953 | 971 | goto err_disable_buffers; |
---|
954 | 972 | } |
---|
955 | 973 | |
---|
956 | | - indio_dev->currentmode = config->mode; |
---|
| 974 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
---|
| 975 | + ret = iio_trigger_attach_poll_func(indio_dev->trig, |
---|
| 976 | + indio_dev->pollfunc); |
---|
| 977 | + if (ret) |
---|
| 978 | + goto err_disable_buffers; |
---|
| 979 | + } |
---|
957 | 980 | |
---|
958 | 981 | if (indio_dev->setup_ops->postenable) { |
---|
959 | 982 | ret = indio_dev->setup_ops->postenable(indio_dev); |
---|
960 | 983 | if (ret) { |
---|
961 | 984 | dev_dbg(&indio_dev->dev, |
---|
962 | 985 | "Buffer not started: postenable failed (%d)\n", ret); |
---|
963 | | - goto err_disable_buffers; |
---|
| 986 | + goto err_detach_pollfunc; |
---|
964 | 987 | } |
---|
965 | 988 | } |
---|
966 | 989 | |
---|
967 | 990 | return 0; |
---|
968 | 991 | |
---|
| 992 | +err_detach_pollfunc: |
---|
| 993 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
---|
| 994 | + iio_trigger_detach_poll_func(indio_dev->trig, |
---|
| 995 | + indio_dev->pollfunc); |
---|
| 996 | + } |
---|
969 | 997 | err_disable_buffers: |
---|
970 | | - list_for_each_entry_continue_reverse(buffer, &indio_dev->buffer_list, |
---|
| 998 | + list_for_each_entry_continue_reverse(buffer, &iio_dev_opaque->buffer_list, |
---|
971 | 999 | buffer_list) |
---|
972 | 1000 | iio_buffer_disable(buffer, indio_dev); |
---|
973 | 1001 | err_run_postdisable: |
---|
974 | | - indio_dev->currentmode = INDIO_DIRECT_MODE; |
---|
975 | 1002 | if (indio_dev->setup_ops->postdisable) |
---|
976 | 1003 | indio_dev->setup_ops->postdisable(indio_dev); |
---|
977 | 1004 | err_undo_config: |
---|
| 1005 | + indio_dev->currentmode = INDIO_DIRECT_MODE; |
---|
978 | 1006 | indio_dev->active_scan_mask = NULL; |
---|
979 | 1007 | |
---|
980 | 1008 | return ret; |
---|
.. | .. |
---|
982 | 1010 | |
---|
983 | 1011 | static int iio_disable_buffers(struct iio_dev *indio_dev) |
---|
984 | 1012 | { |
---|
| 1013 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
985 | 1014 | struct iio_buffer *buffer; |
---|
986 | 1015 | int ret = 0; |
---|
987 | 1016 | int ret2; |
---|
988 | 1017 | |
---|
989 | 1018 | /* Wind down existing buffers - iff there are any */ |
---|
990 | | - if (list_empty(&indio_dev->buffer_list)) |
---|
| 1019 | + if (list_empty(&iio_dev_opaque->buffer_list)) |
---|
991 | 1020 | return 0; |
---|
992 | 1021 | |
---|
993 | 1022 | /* |
---|
.. | .. |
---|
1003 | 1032 | ret = ret2; |
---|
1004 | 1033 | } |
---|
1005 | 1034 | |
---|
1006 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
---|
| 1035 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
---|
| 1036 | + iio_trigger_detach_poll_func(indio_dev->trig, |
---|
| 1037 | + indio_dev->pollfunc); |
---|
| 1038 | + } |
---|
| 1039 | + |
---|
| 1040 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
1007 | 1041 | ret2 = iio_buffer_disable(buffer, indio_dev); |
---|
1008 | 1042 | if (ret2 && !ret) |
---|
1009 | 1043 | ret = ret2; |
---|
1010 | 1044 | } |
---|
1011 | | - |
---|
1012 | | - indio_dev->currentmode = INDIO_DIRECT_MODE; |
---|
1013 | 1045 | |
---|
1014 | 1046 | if (indio_dev->setup_ops->postdisable) { |
---|
1015 | 1047 | ret2 = indio_dev->setup_ops->postdisable(indio_dev); |
---|
.. | .. |
---|
1019 | 1051 | |
---|
1020 | 1052 | iio_free_scan_mask(indio_dev, indio_dev->active_scan_mask); |
---|
1021 | 1053 | indio_dev->active_scan_mask = NULL; |
---|
| 1054 | + indio_dev->currentmode = INDIO_DIRECT_MODE; |
---|
1022 | 1055 | |
---|
1023 | 1056 | return ret; |
---|
1024 | 1057 | } |
---|
.. | .. |
---|
1027 | 1060 | struct iio_buffer *insert_buffer, |
---|
1028 | 1061 | struct iio_buffer *remove_buffer) |
---|
1029 | 1062 | { |
---|
| 1063 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
1030 | 1064 | struct iio_device_config new_config; |
---|
1031 | 1065 | int ret; |
---|
1032 | 1066 | |
---|
.. | .. |
---|
1051 | 1085 | iio_buffer_activate(indio_dev, insert_buffer); |
---|
1052 | 1086 | |
---|
1053 | 1087 | /* If no buffers in list, we are done */ |
---|
1054 | | - if (list_empty(&indio_dev->buffer_list)) |
---|
| 1088 | + if (list_empty(&iio_dev_opaque->buffer_list)) |
---|
1055 | 1089 | return 0; |
---|
1056 | 1090 | |
---|
1057 | 1091 | ret = iio_enable_buffers(indio_dev, &new_config); |
---|
.. | .. |
---|
1128 | 1162 | int ret; |
---|
1129 | 1163 | bool requested_state; |
---|
1130 | 1164 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
| 1165 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
1131 | 1166 | bool inlist; |
---|
1132 | 1167 | |
---|
1133 | 1168 | ret = strtobool(buf, &requested_state); |
---|
.. | .. |
---|
1137 | 1172 | mutex_lock(&indio_dev->mlock); |
---|
1138 | 1173 | |
---|
1139 | 1174 | /* Find out if it is in the list */ |
---|
1140 | | - inlist = iio_buffer_is_active(indio_dev->buffer); |
---|
| 1175 | + inlist = iio_buffer_is_active(buffer); |
---|
1141 | 1176 | /* Already in desired state */ |
---|
1142 | 1177 | if (inlist == requested_state) |
---|
1143 | 1178 | goto done; |
---|
1144 | 1179 | |
---|
1145 | 1180 | if (requested_state) |
---|
1146 | | - ret = __iio_update_buffers(indio_dev, |
---|
1147 | | - indio_dev->buffer, NULL); |
---|
| 1181 | + ret = __iio_update_buffers(indio_dev, buffer, NULL); |
---|
1148 | 1182 | else |
---|
1149 | | - ret = __iio_update_buffers(indio_dev, |
---|
1150 | | - NULL, indio_dev->buffer); |
---|
| 1183 | + ret = __iio_update_buffers(indio_dev, NULL, buffer); |
---|
1151 | 1184 | |
---|
1152 | 1185 | done: |
---|
1153 | 1186 | mutex_unlock(&indio_dev->mlock); |
---|
.. | .. |
---|
1189 | 1222 | goto out; |
---|
1190 | 1223 | } |
---|
1191 | 1224 | |
---|
1192 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
---|
| 1225 | + if (iio_buffer_is_active(buffer)) { |
---|
1193 | 1226 | ret = -EBUSY; |
---|
1194 | 1227 | goto out; |
---|
1195 | 1228 | } |
---|
.. | .. |
---|
1206 | 1239 | char *buf) |
---|
1207 | 1240 | { |
---|
1208 | 1241 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
---|
1209 | | - size_t bytes; |
---|
| 1242 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
1210 | 1243 | |
---|
1211 | | - bytes = iio_buffer_data_available(indio_dev->buffer); |
---|
1212 | | - |
---|
1213 | | - return sprintf(buf, "%zu\n", bytes); |
---|
| 1244 | + return sprintf(buf, "%zu\n", iio_buffer_data_available(buffer)); |
---|
1214 | 1245 | } |
---|
1215 | 1246 | |
---|
1216 | 1247 | static DEVICE_ATTR(length, S_IRUGO | S_IWUSR, iio_buffer_read_length, |
---|
.. | .. |
---|
1233 | 1264 | &dev_attr_data_available.attr, |
---|
1234 | 1265 | }; |
---|
1235 | 1266 | |
---|
1236 | | -int iio_buffer_alloc_sysfs_and_mask(struct iio_dev *indio_dev) |
---|
| 1267 | +static int __iio_buffer_alloc_sysfs_and_mask(struct iio_buffer *buffer, |
---|
| 1268 | + struct iio_dev *indio_dev) |
---|
1237 | 1269 | { |
---|
1238 | 1270 | struct iio_dev_attr *p; |
---|
1239 | 1271 | struct attribute **attr; |
---|
1240 | | - struct iio_buffer *buffer = indio_dev->buffer; |
---|
1241 | | - int ret, i, attrn, attrcount, attrcount_orig = 0; |
---|
| 1272 | + int ret, i, attrn, attrcount; |
---|
1242 | 1273 | const struct iio_chan_spec *channels; |
---|
1243 | | - |
---|
1244 | | - channels = indio_dev->channels; |
---|
1245 | | - if (channels) { |
---|
1246 | | - int ml = indio_dev->masklength; |
---|
1247 | | - |
---|
1248 | | - for (i = 0; i < indio_dev->num_channels; i++) |
---|
1249 | | - ml = max(ml, channels[i].scan_index + 1); |
---|
1250 | | - indio_dev->masklength = ml; |
---|
1251 | | - } |
---|
1252 | | - |
---|
1253 | | - if (!buffer) |
---|
1254 | | - return 0; |
---|
1255 | 1274 | |
---|
1256 | 1275 | attrcount = 0; |
---|
1257 | 1276 | if (buffer->attrs) { |
---|
.. | .. |
---|
1282 | 1301 | |
---|
1283 | 1302 | indio_dev->groups[indio_dev->groupcounter++] = &buffer->buffer_group; |
---|
1284 | 1303 | |
---|
1285 | | - if (buffer->scan_el_attrs != NULL) { |
---|
1286 | | - attr = buffer->scan_el_attrs->attrs; |
---|
1287 | | - while (*attr++ != NULL) |
---|
1288 | | - attrcount_orig++; |
---|
1289 | | - } |
---|
1290 | | - attrcount = attrcount_orig; |
---|
| 1304 | + attrcount = 0; |
---|
1291 | 1305 | INIT_LIST_HEAD(&buffer->scan_el_dev_attr_list); |
---|
1292 | 1306 | channels = indio_dev->channels; |
---|
1293 | 1307 | if (channels) { |
---|
.. | .. |
---|
1296 | 1310 | if (channels[i].scan_index < 0) |
---|
1297 | 1311 | continue; |
---|
1298 | 1312 | |
---|
1299 | | - ret = iio_buffer_add_channel_sysfs(indio_dev, |
---|
| 1313 | + ret = iio_buffer_add_channel_sysfs(indio_dev, buffer, |
---|
1300 | 1314 | &channels[i]); |
---|
1301 | 1315 | if (ret < 0) |
---|
1302 | 1316 | goto error_cleanup_dynamic; |
---|
.. | .. |
---|
1306 | 1320 | channels[i].scan_index; |
---|
1307 | 1321 | } |
---|
1308 | 1322 | if (indio_dev->masklength && buffer->scan_mask == NULL) { |
---|
1309 | | - buffer->scan_mask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
---|
1310 | | - sizeof(*buffer->scan_mask), |
---|
1311 | | - GFP_KERNEL); |
---|
| 1323 | + buffer->scan_mask = bitmap_zalloc(indio_dev->masklength, |
---|
| 1324 | + GFP_KERNEL); |
---|
1312 | 1325 | if (buffer->scan_mask == NULL) { |
---|
1313 | 1326 | ret = -ENOMEM; |
---|
1314 | 1327 | goto error_cleanup_dynamic; |
---|
.. | .. |
---|
1325 | 1338 | ret = -ENOMEM; |
---|
1326 | 1339 | goto error_free_scan_mask; |
---|
1327 | 1340 | } |
---|
1328 | | - if (buffer->scan_el_attrs) |
---|
1329 | | - memcpy(buffer->scan_el_group.attrs, buffer->scan_el_attrs, |
---|
1330 | | - sizeof(buffer->scan_el_group.attrs[0])*attrcount_orig); |
---|
1331 | | - attrn = attrcount_orig; |
---|
| 1341 | + attrn = 0; |
---|
1332 | 1342 | |
---|
1333 | 1343 | list_for_each_entry(p, &buffer->scan_el_dev_attr_list, l) |
---|
1334 | 1344 | buffer->scan_el_group.attrs[attrn++] = &p->dev_attr.attr; |
---|
.. | .. |
---|
1337 | 1347 | return 0; |
---|
1338 | 1348 | |
---|
1339 | 1349 | error_free_scan_mask: |
---|
1340 | | - kfree(buffer->scan_mask); |
---|
| 1350 | + bitmap_free(buffer->scan_mask); |
---|
1341 | 1351 | error_cleanup_dynamic: |
---|
1342 | 1352 | iio_free_chan_devattr_list(&buffer->scan_el_dev_attr_list); |
---|
1343 | | - kfree(indio_dev->buffer->buffer_group.attrs); |
---|
| 1353 | + kfree(buffer->buffer_group.attrs); |
---|
1344 | 1354 | |
---|
1345 | 1355 | return ret; |
---|
1346 | 1356 | } |
---|
1347 | 1357 | |
---|
| 1358 | +int iio_buffer_alloc_sysfs_and_mask(struct iio_dev *indio_dev) |
---|
| 1359 | +{ |
---|
| 1360 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
| 1361 | + const struct iio_chan_spec *channels; |
---|
| 1362 | + int i; |
---|
| 1363 | + |
---|
| 1364 | + channels = indio_dev->channels; |
---|
| 1365 | + if (channels) { |
---|
| 1366 | + int ml = indio_dev->masklength; |
---|
| 1367 | + |
---|
| 1368 | + for (i = 0; i < indio_dev->num_channels; i++) |
---|
| 1369 | + ml = max(ml, channels[i].scan_index + 1); |
---|
| 1370 | + indio_dev->masklength = ml; |
---|
| 1371 | + } |
---|
| 1372 | + |
---|
| 1373 | + if (!buffer) |
---|
| 1374 | + return 0; |
---|
| 1375 | + |
---|
| 1376 | + return __iio_buffer_alloc_sysfs_and_mask(buffer, indio_dev); |
---|
| 1377 | +} |
---|
| 1378 | + |
---|
| 1379 | +static void __iio_buffer_free_sysfs_and_mask(struct iio_buffer *buffer) |
---|
| 1380 | +{ |
---|
| 1381 | + bitmap_free(buffer->scan_mask); |
---|
| 1382 | + kfree(buffer->buffer_group.attrs); |
---|
| 1383 | + kfree(buffer->scan_el_group.attrs); |
---|
| 1384 | + iio_free_chan_devattr_list(&buffer->scan_el_dev_attr_list); |
---|
| 1385 | +} |
---|
| 1386 | + |
---|
1348 | 1387 | void iio_buffer_free_sysfs_and_mask(struct iio_dev *indio_dev) |
---|
1349 | 1388 | { |
---|
1350 | | - if (!indio_dev->buffer) |
---|
| 1389 | + struct iio_buffer *buffer = indio_dev->buffer; |
---|
| 1390 | + |
---|
| 1391 | + if (!buffer) |
---|
1351 | 1392 | return; |
---|
1352 | 1393 | |
---|
1353 | | - kfree(indio_dev->buffer->scan_mask); |
---|
1354 | | - kfree(indio_dev->buffer->buffer_group.attrs); |
---|
1355 | | - kfree(indio_dev->buffer->scan_el_group.attrs); |
---|
1356 | | - iio_free_chan_devattr_list(&indio_dev->buffer->scan_el_dev_attr_list); |
---|
| 1394 | + __iio_buffer_free_sysfs_and_mask(buffer); |
---|
1357 | 1395 | } |
---|
1358 | 1396 | |
---|
1359 | 1397 | /** |
---|
.. | .. |
---|
1410 | 1448 | */ |
---|
1411 | 1449 | int iio_push_to_buffers(struct iio_dev *indio_dev, const void *data) |
---|
1412 | 1450 | { |
---|
| 1451 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
---|
1413 | 1452 | int ret; |
---|
1414 | 1453 | struct iio_buffer *buf; |
---|
1415 | 1454 | |
---|
1416 | | - list_for_each_entry(buf, &indio_dev->buffer_list, buffer_list) { |
---|
| 1455 | + list_for_each_entry(buf, &iio_dev_opaque->buffer_list, buffer_list) { |
---|
1417 | 1456 | ret = iio_push_to_buffer(buf, data); |
---|
1418 | 1457 | if (ret < 0) |
---|
1419 | 1458 | return ret; |
---|