| .. | .. |
|---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-only |
|---|
| 1 | 2 | /* The industrial I/O core |
|---|
| 2 | 3 | * |
|---|
| 3 | 4 | * Copyright (c) 2008 Jonathan Cameron |
|---|
| 4 | 5 | * |
|---|
| 5 | | - * This program is free software; you can redistribute it and/or modify it |
|---|
| 6 | | - * under the terms of the GNU General Public License version 2 as published by |
|---|
| 7 | | - * the Free Software Foundation. |
|---|
| 8 | | - * |
|---|
| 9 | 6 | * Handling of buffer allocation / resizing. |
|---|
| 10 | | - * |
|---|
| 11 | 7 | * |
|---|
| 12 | 8 | * Things to look at here. |
|---|
| 13 | 9 | * - Better memory allocation techniques? |
|---|
| .. | .. |
|---|
| 23 | 19 | #include <linux/sched/signal.h> |
|---|
| 24 | 20 | |
|---|
| 25 | 21 | #include <linux/iio/iio.h> |
|---|
| 22 | +#include <linux/iio/iio-opaque.h> |
|---|
| 26 | 23 | #include "iio_core.h" |
|---|
| 24 | +#include "iio_core_trigger.h" |
|---|
| 27 | 25 | #include <linux/iio/sysfs.h> |
|---|
| 28 | 26 | #include <linux/iio/buffer.h> |
|---|
| 29 | 27 | #include <linux/iio/buffer_impl.h> |
|---|
| .. | .. |
|---|
| 91 | 89 | } |
|---|
| 92 | 90 | |
|---|
| 93 | 91 | /** |
|---|
| 94 | | - * iio_buffer_read_first_n_outer() - chrdev read for buffer access |
|---|
| 92 | + * iio_buffer_read_outer() - chrdev read for buffer access |
|---|
| 95 | 93 | * @filp: File structure pointer for the char device |
|---|
| 96 | 94 | * @buf: Destination buffer for iio buffer read |
|---|
| 97 | 95 | * @n: First n bytes to read |
|---|
| .. | .. |
|---|
| 103 | 101 | * Return: negative values corresponding to error codes or ret != 0 |
|---|
| 104 | 102 | * for ending the reading activity |
|---|
| 105 | 103 | **/ |
|---|
| 106 | | -ssize_t iio_buffer_read_first_n_outer(struct file *filp, char __user *buf, |
|---|
| 107 | | - size_t n, loff_t *f_ps) |
|---|
| 104 | +ssize_t iio_buffer_read_outer(struct file *filp, char __user *buf, |
|---|
| 105 | + size_t n, loff_t *f_ps) |
|---|
| 108 | 106 | { |
|---|
| 109 | 107 | struct iio_dev *indio_dev = filp->private_data; |
|---|
| 110 | 108 | struct iio_buffer *rb = indio_dev->buffer; |
|---|
| .. | .. |
|---|
| 116 | 114 | if (!indio_dev->info) |
|---|
| 117 | 115 | return -ENODEV; |
|---|
| 118 | 116 | |
|---|
| 119 | | - if (!rb || !rb->access->read_first_n) |
|---|
| 117 | + if (!rb || !rb->access->read) |
|---|
| 120 | 118 | return -EINVAL; |
|---|
| 121 | 119 | |
|---|
| 122 | 120 | datum_size = rb->bytes_per_datum; |
|---|
| .. | .. |
|---|
| 151 | 149 | continue; |
|---|
| 152 | 150 | } |
|---|
| 153 | 151 | |
|---|
| 154 | | - ret = rb->access->read_first_n(rb, n, buf); |
|---|
| 152 | + ret = rb->access->read(rb, n, buf); |
|---|
| 155 | 153 | if (ret == 0 && (filp->f_flags & O_NONBLOCK)) |
|---|
| 156 | 154 | ret = -EAGAIN; |
|---|
| 157 | 155 | } while (ret == 0); |
|---|
| .. | .. |
|---|
| 193 | 191 | */ |
|---|
| 194 | 192 | void iio_buffer_wakeup_poll(struct iio_dev *indio_dev) |
|---|
| 195 | 193 | { |
|---|
| 196 | | - if (!indio_dev->buffer) |
|---|
| 194 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 195 | + |
|---|
| 196 | + if (!buffer) |
|---|
| 197 | 197 | return; |
|---|
| 198 | 198 | |
|---|
| 199 | | - wake_up(&indio_dev->buffer->pollq); |
|---|
| 199 | + wake_up(&buffer->pollq); |
|---|
| 200 | 200 | } |
|---|
| 201 | 201 | |
|---|
| 202 | 202 | void iio_buffer_init(struct iio_buffer *buffer) |
|---|
| .. | .. |
|---|
| 266 | 266 | { |
|---|
| 267 | 267 | int ret; |
|---|
| 268 | 268 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 269 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 269 | 270 | |
|---|
| 270 | 271 | /* Ensure ret is 0 or 1. */ |
|---|
| 271 | 272 | ret = !!test_bit(to_iio_dev_attr(attr)->address, |
|---|
| 272 | | - indio_dev->buffer->scan_mask); |
|---|
| 273 | + buffer->scan_mask); |
|---|
| 273 | 274 | |
|---|
| 274 | 275 | return sprintf(buf, "%d\n", ret); |
|---|
| 275 | 276 | } |
|---|
| .. | .. |
|---|
| 320 | 321 | const unsigned long *mask; |
|---|
| 321 | 322 | unsigned long *trialmask; |
|---|
| 322 | 323 | |
|---|
| 323 | | - trialmask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
|---|
| 324 | | - sizeof(*trialmask), GFP_KERNEL); |
|---|
| 324 | + trialmask = bitmap_zalloc(indio_dev->masklength, GFP_KERNEL); |
|---|
| 325 | 325 | if (trialmask == NULL) |
|---|
| 326 | 326 | return -ENOMEM; |
|---|
| 327 | 327 | if (!indio_dev->masklength) { |
|---|
| .. | .. |
|---|
| 343 | 343 | } |
|---|
| 344 | 344 | bitmap_copy(buffer->scan_mask, trialmask, indio_dev->masklength); |
|---|
| 345 | 345 | |
|---|
| 346 | | - kfree(trialmask); |
|---|
| 346 | + bitmap_free(trialmask); |
|---|
| 347 | 347 | |
|---|
| 348 | 348 | return 0; |
|---|
| 349 | 349 | |
|---|
| 350 | 350 | err_invalid_mask: |
|---|
| 351 | | - kfree(trialmask); |
|---|
| 351 | + bitmap_free(trialmask); |
|---|
| 352 | 352 | return -EINVAL; |
|---|
| 353 | 353 | } |
|---|
| 354 | 354 | |
|---|
| .. | .. |
|---|
| 386 | 386 | if (ret < 0) |
|---|
| 387 | 387 | return ret; |
|---|
| 388 | 388 | mutex_lock(&indio_dev->mlock); |
|---|
| 389 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
|---|
| 389 | + if (iio_buffer_is_active(buffer)) { |
|---|
| 390 | 390 | ret = -EBUSY; |
|---|
| 391 | 391 | goto error_ret; |
|---|
| 392 | 392 | } |
|---|
| .. | .. |
|---|
| 415 | 415 | char *buf) |
|---|
| 416 | 416 | { |
|---|
| 417 | 417 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 418 | | - return sprintf(buf, "%d\n", indio_dev->buffer->scan_timestamp); |
|---|
| 418 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 419 | + |
|---|
| 420 | + return sprintf(buf, "%d\n", buffer->scan_timestamp); |
|---|
| 419 | 421 | } |
|---|
| 420 | 422 | |
|---|
| 421 | 423 | static ssize_t iio_scan_el_ts_store(struct device *dev, |
|---|
| .. | .. |
|---|
| 425 | 427 | { |
|---|
| 426 | 428 | int ret; |
|---|
| 427 | 429 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 430 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 428 | 431 | bool state; |
|---|
| 429 | 432 | |
|---|
| 430 | 433 | ret = strtobool(buf, &state); |
|---|
| .. | .. |
|---|
| 432 | 435 | return ret; |
|---|
| 433 | 436 | |
|---|
| 434 | 437 | mutex_lock(&indio_dev->mlock); |
|---|
| 435 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
|---|
| 438 | + if (iio_buffer_is_active(buffer)) { |
|---|
| 436 | 439 | ret = -EBUSY; |
|---|
| 437 | 440 | goto error_ret; |
|---|
| 438 | 441 | } |
|---|
| 439 | | - indio_dev->buffer->scan_timestamp = state; |
|---|
| 442 | + buffer->scan_timestamp = state; |
|---|
| 440 | 443 | error_ret: |
|---|
| 441 | 444 | mutex_unlock(&indio_dev->mlock); |
|---|
| 442 | 445 | |
|---|
| .. | .. |
|---|
| 444 | 447 | } |
|---|
| 445 | 448 | |
|---|
| 446 | 449 | static int iio_buffer_add_channel_sysfs(struct iio_dev *indio_dev, |
|---|
| 450 | + struct iio_buffer *buffer, |
|---|
| 447 | 451 | const struct iio_chan_spec *chan) |
|---|
| 448 | 452 | { |
|---|
| 449 | 453 | int ret, attrcount = 0; |
|---|
| 450 | | - struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 451 | 454 | |
|---|
| 452 | 455 | ret = __iio_add_chan_devattr("index", |
|---|
| 453 | 456 | chan, |
|---|
| .. | .. |
|---|
| 523 | 526 | return len; |
|---|
| 524 | 527 | |
|---|
| 525 | 528 | mutex_lock(&indio_dev->mlock); |
|---|
| 526 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
|---|
| 529 | + if (iio_buffer_is_active(buffer)) { |
|---|
| 527 | 530 | ret = -EBUSY; |
|---|
| 528 | 531 | } else { |
|---|
| 529 | 532 | buffer->access->set_length(buffer, val); |
|---|
| .. | .. |
|---|
| 544 | 547 | char *buf) |
|---|
| 545 | 548 | { |
|---|
| 546 | 549 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 547 | | - return sprintf(buf, "%d\n", iio_buffer_is_active(indio_dev->buffer)); |
|---|
| 550 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 551 | + |
|---|
| 552 | + return sprintf(buf, "%d\n", iio_buffer_is_active(buffer)); |
|---|
| 548 | 553 | } |
|---|
| 549 | 554 | |
|---|
| 550 | 555 | static unsigned int iio_storage_bytes_for_si(struct iio_dev *indio_dev, |
|---|
| .. | .. |
|---|
| 595 | 600 | static void iio_buffer_activate(struct iio_dev *indio_dev, |
|---|
| 596 | 601 | struct iio_buffer *buffer) |
|---|
| 597 | 602 | { |
|---|
| 603 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 604 | + |
|---|
| 598 | 605 | iio_buffer_get(buffer); |
|---|
| 599 | | - list_add(&buffer->buffer_list, &indio_dev->buffer_list); |
|---|
| 606 | + list_add(&buffer->buffer_list, &iio_dev_opaque->buffer_list); |
|---|
| 600 | 607 | } |
|---|
| 601 | 608 | |
|---|
| 602 | 609 | static void iio_buffer_deactivate(struct iio_buffer *buffer) |
|---|
| .. | .. |
|---|
| 608 | 615 | |
|---|
| 609 | 616 | static void iio_buffer_deactivate_all(struct iio_dev *indio_dev) |
|---|
| 610 | 617 | { |
|---|
| 618 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 611 | 619 | struct iio_buffer *buffer, *_buffer; |
|---|
| 612 | 620 | |
|---|
| 613 | 621 | list_for_each_entry_safe(buffer, _buffer, |
|---|
| 614 | | - &indio_dev->buffer_list, buffer_list) |
|---|
| 622 | + &iio_dev_opaque->buffer_list, buffer_list) |
|---|
| 615 | 623 | iio_buffer_deactivate(buffer); |
|---|
| 616 | 624 | } |
|---|
| 617 | 625 | |
|---|
| .. | .. |
|---|
| 669 | 677 | { |
|---|
| 670 | 678 | /* If the mask is dynamically allocated free it, otherwise do nothing */ |
|---|
| 671 | 679 | if (!indio_dev->available_scan_masks) |
|---|
| 672 | | - kfree(mask); |
|---|
| 680 | + bitmap_free(mask); |
|---|
| 673 | 681 | } |
|---|
| 674 | 682 | |
|---|
| 675 | 683 | struct iio_device_config { |
|---|
| .. | .. |
|---|
| 684 | 692 | struct iio_buffer *insert_buffer, struct iio_buffer *remove_buffer, |
|---|
| 685 | 693 | struct iio_device_config *config) |
|---|
| 686 | 694 | { |
|---|
| 695 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 687 | 696 | unsigned long *compound_mask; |
|---|
| 688 | 697 | const unsigned long *scan_mask; |
|---|
| 689 | 698 | bool strict_scanmask = false; |
|---|
| 690 | 699 | struct iio_buffer *buffer; |
|---|
| 691 | 700 | bool scan_timestamp; |
|---|
| 692 | 701 | unsigned int modes; |
|---|
| 702 | + |
|---|
| 703 | + if (insert_buffer && |
|---|
| 704 | + bitmap_empty(insert_buffer->scan_mask, indio_dev->masklength)) { |
|---|
| 705 | + dev_dbg(&indio_dev->dev, |
|---|
| 706 | + "At least one scan element must be enabled first\n"); |
|---|
| 707 | + return -EINVAL; |
|---|
| 708 | + } |
|---|
| 693 | 709 | |
|---|
| 694 | 710 | memset(config, 0, sizeof(*config)); |
|---|
| 695 | 711 | config->watermark = ~0; |
|---|
| .. | .. |
|---|
| 699 | 715 | * to verify. |
|---|
| 700 | 716 | */ |
|---|
| 701 | 717 | if (remove_buffer && !insert_buffer && |
|---|
| 702 | | - list_is_singular(&indio_dev->buffer_list)) |
|---|
| 718 | + list_is_singular(&iio_dev_opaque->buffer_list)) |
|---|
| 703 | 719 | return 0; |
|---|
| 704 | 720 | |
|---|
| 705 | 721 | modes = indio_dev->modes; |
|---|
| 706 | 722 | |
|---|
| 707 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
|---|
| 723 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 708 | 724 | if (buffer == remove_buffer) |
|---|
| 709 | 725 | continue; |
|---|
| 710 | 726 | modes &= buffer->access->modes; |
|---|
| .. | .. |
|---|
| 725 | 741 | * Keep things simple for now and only allow a single buffer to |
|---|
| 726 | 742 | * be connected in hardware mode. |
|---|
| 727 | 743 | */ |
|---|
| 728 | | - if (insert_buffer && !list_empty(&indio_dev->buffer_list)) |
|---|
| 744 | + if (insert_buffer && !list_empty(&iio_dev_opaque->buffer_list)) |
|---|
| 729 | 745 | return -EINVAL; |
|---|
| 730 | 746 | config->mode = INDIO_BUFFER_HARDWARE; |
|---|
| 731 | 747 | strict_scanmask = true; |
|---|
| .. | .. |
|---|
| 739 | 755 | } |
|---|
| 740 | 756 | |
|---|
| 741 | 757 | /* What scan mask do we actually have? */ |
|---|
| 742 | | - compound_mask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
|---|
| 743 | | - sizeof(long), GFP_KERNEL); |
|---|
| 758 | + compound_mask = bitmap_zalloc(indio_dev->masklength, GFP_KERNEL); |
|---|
| 744 | 759 | if (compound_mask == NULL) |
|---|
| 745 | 760 | return -ENOMEM; |
|---|
| 746 | 761 | |
|---|
| 747 | 762 | scan_timestamp = false; |
|---|
| 748 | 763 | |
|---|
| 749 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
|---|
| 764 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 750 | 765 | if (buffer == remove_buffer) |
|---|
| 751 | 766 | continue; |
|---|
| 752 | 767 | bitmap_or(compound_mask, compound_mask, buffer->scan_mask, |
|---|
| .. | .. |
|---|
| 765 | 780 | indio_dev->masklength, |
|---|
| 766 | 781 | compound_mask, |
|---|
| 767 | 782 | strict_scanmask); |
|---|
| 768 | | - kfree(compound_mask); |
|---|
| 783 | + bitmap_free(compound_mask); |
|---|
| 769 | 784 | if (scan_mask == NULL) |
|---|
| 770 | 785 | return -EINVAL; |
|---|
| 771 | 786 | } else { |
|---|
| .. | .. |
|---|
| 892 | 907 | |
|---|
| 893 | 908 | static int iio_update_demux(struct iio_dev *indio_dev) |
|---|
| 894 | 909 | { |
|---|
| 910 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 895 | 911 | struct iio_buffer *buffer; |
|---|
| 896 | 912 | int ret; |
|---|
| 897 | 913 | |
|---|
| 898 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
|---|
| 914 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 899 | 915 | ret = iio_buffer_update_demux(indio_dev, buffer); |
|---|
| 900 | 916 | if (ret < 0) |
|---|
| 901 | 917 | goto error_clear_mux_table; |
|---|
| .. | .. |
|---|
| 903 | 919 | return 0; |
|---|
| 904 | 920 | |
|---|
| 905 | 921 | error_clear_mux_table: |
|---|
| 906 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) |
|---|
| 922 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) |
|---|
| 907 | 923 | iio_buffer_demux_free(buffer); |
|---|
| 908 | 924 | |
|---|
| 909 | 925 | return ret; |
|---|
| .. | .. |
|---|
| 912 | 928 | static int iio_enable_buffers(struct iio_dev *indio_dev, |
|---|
| 913 | 929 | struct iio_device_config *config) |
|---|
| 914 | 930 | { |
|---|
| 931 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 915 | 932 | struct iio_buffer *buffer; |
|---|
| 916 | 933 | int ret; |
|---|
| 917 | 934 | |
|---|
| 918 | 935 | indio_dev->active_scan_mask = config->scan_mask; |
|---|
| 919 | 936 | indio_dev->scan_timestamp = config->scan_timestamp; |
|---|
| 920 | 937 | indio_dev->scan_bytes = config->scan_bytes; |
|---|
| 938 | + indio_dev->currentmode = config->mode; |
|---|
| 921 | 939 | |
|---|
| 922 | 940 | iio_update_demux(indio_dev); |
|---|
| 923 | 941 | |
|---|
| .. | .. |
|---|
| 947 | 965 | indio_dev->info->hwfifo_set_watermark(indio_dev, |
|---|
| 948 | 966 | config->watermark); |
|---|
| 949 | 967 | |
|---|
| 950 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
|---|
| 968 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 951 | 969 | ret = iio_buffer_enable(buffer, indio_dev); |
|---|
| 952 | 970 | if (ret) |
|---|
| 953 | 971 | goto err_disable_buffers; |
|---|
| 954 | 972 | } |
|---|
| 955 | 973 | |
|---|
| 956 | | - indio_dev->currentmode = config->mode; |
|---|
| 974 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
|---|
| 975 | + ret = iio_trigger_attach_poll_func(indio_dev->trig, |
|---|
| 976 | + indio_dev->pollfunc); |
|---|
| 977 | + if (ret) |
|---|
| 978 | + goto err_disable_buffers; |
|---|
| 979 | + } |
|---|
| 957 | 980 | |
|---|
| 958 | 981 | if (indio_dev->setup_ops->postenable) { |
|---|
| 959 | 982 | ret = indio_dev->setup_ops->postenable(indio_dev); |
|---|
| 960 | 983 | if (ret) { |
|---|
| 961 | 984 | dev_dbg(&indio_dev->dev, |
|---|
| 962 | 985 | "Buffer not started: postenable failed (%d)\n", ret); |
|---|
| 963 | | - goto err_disable_buffers; |
|---|
| 986 | + goto err_detach_pollfunc; |
|---|
| 964 | 987 | } |
|---|
| 965 | 988 | } |
|---|
| 966 | 989 | |
|---|
| 967 | 990 | return 0; |
|---|
| 968 | 991 | |
|---|
| 992 | +err_detach_pollfunc: |
|---|
| 993 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
|---|
| 994 | + iio_trigger_detach_poll_func(indio_dev->trig, |
|---|
| 995 | + indio_dev->pollfunc); |
|---|
| 996 | + } |
|---|
| 969 | 997 | err_disable_buffers: |
|---|
| 970 | | - list_for_each_entry_continue_reverse(buffer, &indio_dev->buffer_list, |
|---|
| 998 | + list_for_each_entry_continue_reverse(buffer, &iio_dev_opaque->buffer_list, |
|---|
| 971 | 999 | buffer_list) |
|---|
| 972 | 1000 | iio_buffer_disable(buffer, indio_dev); |
|---|
| 973 | 1001 | err_run_postdisable: |
|---|
| 974 | | - indio_dev->currentmode = INDIO_DIRECT_MODE; |
|---|
| 975 | 1002 | if (indio_dev->setup_ops->postdisable) |
|---|
| 976 | 1003 | indio_dev->setup_ops->postdisable(indio_dev); |
|---|
| 977 | 1004 | err_undo_config: |
|---|
| 1005 | + indio_dev->currentmode = INDIO_DIRECT_MODE; |
|---|
| 978 | 1006 | indio_dev->active_scan_mask = NULL; |
|---|
| 979 | 1007 | |
|---|
| 980 | 1008 | return ret; |
|---|
| .. | .. |
|---|
| 982 | 1010 | |
|---|
| 983 | 1011 | static int iio_disable_buffers(struct iio_dev *indio_dev) |
|---|
| 984 | 1012 | { |
|---|
| 1013 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 985 | 1014 | struct iio_buffer *buffer; |
|---|
| 986 | 1015 | int ret = 0; |
|---|
| 987 | 1016 | int ret2; |
|---|
| 988 | 1017 | |
|---|
| 989 | 1018 | /* Wind down existing buffers - iff there are any */ |
|---|
| 990 | | - if (list_empty(&indio_dev->buffer_list)) |
|---|
| 1019 | + if (list_empty(&iio_dev_opaque->buffer_list)) |
|---|
| 991 | 1020 | return 0; |
|---|
| 992 | 1021 | |
|---|
| 993 | 1022 | /* |
|---|
| .. | .. |
|---|
| 1003 | 1032 | ret = ret2; |
|---|
| 1004 | 1033 | } |
|---|
| 1005 | 1034 | |
|---|
| 1006 | | - list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { |
|---|
| 1035 | + if (indio_dev->currentmode == INDIO_BUFFER_TRIGGERED) { |
|---|
| 1036 | + iio_trigger_detach_poll_func(indio_dev->trig, |
|---|
| 1037 | + indio_dev->pollfunc); |
|---|
| 1038 | + } |
|---|
| 1039 | + |
|---|
| 1040 | + list_for_each_entry(buffer, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 1007 | 1041 | ret2 = iio_buffer_disable(buffer, indio_dev); |
|---|
| 1008 | 1042 | if (ret2 && !ret) |
|---|
| 1009 | 1043 | ret = ret2; |
|---|
| 1010 | 1044 | } |
|---|
| 1011 | | - |
|---|
| 1012 | | - indio_dev->currentmode = INDIO_DIRECT_MODE; |
|---|
| 1013 | 1045 | |
|---|
| 1014 | 1046 | if (indio_dev->setup_ops->postdisable) { |
|---|
| 1015 | 1047 | ret2 = indio_dev->setup_ops->postdisable(indio_dev); |
|---|
| .. | .. |
|---|
| 1019 | 1051 | |
|---|
| 1020 | 1052 | iio_free_scan_mask(indio_dev, indio_dev->active_scan_mask); |
|---|
| 1021 | 1053 | indio_dev->active_scan_mask = NULL; |
|---|
| 1054 | + indio_dev->currentmode = INDIO_DIRECT_MODE; |
|---|
| 1022 | 1055 | |
|---|
| 1023 | 1056 | return ret; |
|---|
| 1024 | 1057 | } |
|---|
| .. | .. |
|---|
| 1027 | 1060 | struct iio_buffer *insert_buffer, |
|---|
| 1028 | 1061 | struct iio_buffer *remove_buffer) |
|---|
| 1029 | 1062 | { |
|---|
| 1063 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 1030 | 1064 | struct iio_device_config new_config; |
|---|
| 1031 | 1065 | int ret; |
|---|
| 1032 | 1066 | |
|---|
| .. | .. |
|---|
| 1051 | 1085 | iio_buffer_activate(indio_dev, insert_buffer); |
|---|
| 1052 | 1086 | |
|---|
| 1053 | 1087 | /* If no buffers in list, we are done */ |
|---|
| 1054 | | - if (list_empty(&indio_dev->buffer_list)) |
|---|
| 1088 | + if (list_empty(&iio_dev_opaque->buffer_list)) |
|---|
| 1055 | 1089 | return 0; |
|---|
| 1056 | 1090 | |
|---|
| 1057 | 1091 | ret = iio_enable_buffers(indio_dev, &new_config); |
|---|
| .. | .. |
|---|
| 1128 | 1162 | int ret; |
|---|
| 1129 | 1163 | bool requested_state; |
|---|
| 1130 | 1164 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 1165 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 1131 | 1166 | bool inlist; |
|---|
| 1132 | 1167 | |
|---|
| 1133 | 1168 | ret = strtobool(buf, &requested_state); |
|---|
| .. | .. |
|---|
| 1137 | 1172 | mutex_lock(&indio_dev->mlock); |
|---|
| 1138 | 1173 | |
|---|
| 1139 | 1174 | /* Find out if it is in the list */ |
|---|
| 1140 | | - inlist = iio_buffer_is_active(indio_dev->buffer); |
|---|
| 1175 | + inlist = iio_buffer_is_active(buffer); |
|---|
| 1141 | 1176 | /* Already in desired state */ |
|---|
| 1142 | 1177 | if (inlist == requested_state) |
|---|
| 1143 | 1178 | goto done; |
|---|
| 1144 | 1179 | |
|---|
| 1145 | 1180 | if (requested_state) |
|---|
| 1146 | | - ret = __iio_update_buffers(indio_dev, |
|---|
| 1147 | | - indio_dev->buffer, NULL); |
|---|
| 1181 | + ret = __iio_update_buffers(indio_dev, buffer, NULL); |
|---|
| 1148 | 1182 | else |
|---|
| 1149 | | - ret = __iio_update_buffers(indio_dev, |
|---|
| 1150 | | - NULL, indio_dev->buffer); |
|---|
| 1183 | + ret = __iio_update_buffers(indio_dev, NULL, buffer); |
|---|
| 1151 | 1184 | |
|---|
| 1152 | 1185 | done: |
|---|
| 1153 | 1186 | mutex_unlock(&indio_dev->mlock); |
|---|
| .. | .. |
|---|
| 1189 | 1222 | goto out; |
|---|
| 1190 | 1223 | } |
|---|
| 1191 | 1224 | |
|---|
| 1192 | | - if (iio_buffer_is_active(indio_dev->buffer)) { |
|---|
| 1225 | + if (iio_buffer_is_active(buffer)) { |
|---|
| 1193 | 1226 | ret = -EBUSY; |
|---|
| 1194 | 1227 | goto out; |
|---|
| 1195 | 1228 | } |
|---|
| .. | .. |
|---|
| 1206 | 1239 | char *buf) |
|---|
| 1207 | 1240 | { |
|---|
| 1208 | 1241 | struct iio_dev *indio_dev = dev_to_iio_dev(dev); |
|---|
| 1209 | | - size_t bytes; |
|---|
| 1242 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 1210 | 1243 | |
|---|
| 1211 | | - bytes = iio_buffer_data_available(indio_dev->buffer); |
|---|
| 1212 | | - |
|---|
| 1213 | | - return sprintf(buf, "%zu\n", bytes); |
|---|
| 1244 | + return sprintf(buf, "%zu\n", iio_buffer_data_available(buffer)); |
|---|
| 1214 | 1245 | } |
|---|
| 1215 | 1246 | |
|---|
| 1216 | 1247 | static DEVICE_ATTR(length, S_IRUGO | S_IWUSR, iio_buffer_read_length, |
|---|
| .. | .. |
|---|
| 1233 | 1264 | &dev_attr_data_available.attr, |
|---|
| 1234 | 1265 | }; |
|---|
| 1235 | 1266 | |
|---|
| 1236 | | -int iio_buffer_alloc_sysfs_and_mask(struct iio_dev *indio_dev) |
|---|
| 1267 | +static int __iio_buffer_alloc_sysfs_and_mask(struct iio_buffer *buffer, |
|---|
| 1268 | + struct iio_dev *indio_dev) |
|---|
| 1237 | 1269 | { |
|---|
| 1238 | 1270 | struct iio_dev_attr *p; |
|---|
| 1239 | 1271 | struct attribute **attr; |
|---|
| 1240 | | - struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 1241 | | - int ret, i, attrn, attrcount, attrcount_orig = 0; |
|---|
| 1272 | + int ret, i, attrn, attrcount; |
|---|
| 1242 | 1273 | const struct iio_chan_spec *channels; |
|---|
| 1243 | | - |
|---|
| 1244 | | - channels = indio_dev->channels; |
|---|
| 1245 | | - if (channels) { |
|---|
| 1246 | | - int ml = indio_dev->masklength; |
|---|
| 1247 | | - |
|---|
| 1248 | | - for (i = 0; i < indio_dev->num_channels; i++) |
|---|
| 1249 | | - ml = max(ml, channels[i].scan_index + 1); |
|---|
| 1250 | | - indio_dev->masklength = ml; |
|---|
| 1251 | | - } |
|---|
| 1252 | | - |
|---|
| 1253 | | - if (!buffer) |
|---|
| 1254 | | - return 0; |
|---|
| 1255 | 1274 | |
|---|
| 1256 | 1275 | attrcount = 0; |
|---|
| 1257 | 1276 | if (buffer->attrs) { |
|---|
| .. | .. |
|---|
| 1282 | 1301 | |
|---|
| 1283 | 1302 | indio_dev->groups[indio_dev->groupcounter++] = &buffer->buffer_group; |
|---|
| 1284 | 1303 | |
|---|
| 1285 | | - if (buffer->scan_el_attrs != NULL) { |
|---|
| 1286 | | - attr = buffer->scan_el_attrs->attrs; |
|---|
| 1287 | | - while (*attr++ != NULL) |
|---|
| 1288 | | - attrcount_orig++; |
|---|
| 1289 | | - } |
|---|
| 1290 | | - attrcount = attrcount_orig; |
|---|
| 1304 | + attrcount = 0; |
|---|
| 1291 | 1305 | INIT_LIST_HEAD(&buffer->scan_el_dev_attr_list); |
|---|
| 1292 | 1306 | channels = indio_dev->channels; |
|---|
| 1293 | 1307 | if (channels) { |
|---|
| .. | .. |
|---|
| 1296 | 1310 | if (channels[i].scan_index < 0) |
|---|
| 1297 | 1311 | continue; |
|---|
| 1298 | 1312 | |
|---|
| 1299 | | - ret = iio_buffer_add_channel_sysfs(indio_dev, |
|---|
| 1313 | + ret = iio_buffer_add_channel_sysfs(indio_dev, buffer, |
|---|
| 1300 | 1314 | &channels[i]); |
|---|
| 1301 | 1315 | if (ret < 0) |
|---|
| 1302 | 1316 | goto error_cleanup_dynamic; |
|---|
| .. | .. |
|---|
| 1306 | 1320 | channels[i].scan_index; |
|---|
| 1307 | 1321 | } |
|---|
| 1308 | 1322 | if (indio_dev->masklength && buffer->scan_mask == NULL) { |
|---|
| 1309 | | - buffer->scan_mask = kcalloc(BITS_TO_LONGS(indio_dev->masklength), |
|---|
| 1310 | | - sizeof(*buffer->scan_mask), |
|---|
| 1311 | | - GFP_KERNEL); |
|---|
| 1323 | + buffer->scan_mask = bitmap_zalloc(indio_dev->masklength, |
|---|
| 1324 | + GFP_KERNEL); |
|---|
| 1312 | 1325 | if (buffer->scan_mask == NULL) { |
|---|
| 1313 | 1326 | ret = -ENOMEM; |
|---|
| 1314 | 1327 | goto error_cleanup_dynamic; |
|---|
| .. | .. |
|---|
| 1325 | 1338 | ret = -ENOMEM; |
|---|
| 1326 | 1339 | goto error_free_scan_mask; |
|---|
| 1327 | 1340 | } |
|---|
| 1328 | | - if (buffer->scan_el_attrs) |
|---|
| 1329 | | - memcpy(buffer->scan_el_group.attrs, buffer->scan_el_attrs, |
|---|
| 1330 | | - sizeof(buffer->scan_el_group.attrs[0])*attrcount_orig); |
|---|
| 1331 | | - attrn = attrcount_orig; |
|---|
| 1341 | + attrn = 0; |
|---|
| 1332 | 1342 | |
|---|
| 1333 | 1343 | list_for_each_entry(p, &buffer->scan_el_dev_attr_list, l) |
|---|
| 1334 | 1344 | buffer->scan_el_group.attrs[attrn++] = &p->dev_attr.attr; |
|---|
| .. | .. |
|---|
| 1337 | 1347 | return 0; |
|---|
| 1338 | 1348 | |
|---|
| 1339 | 1349 | error_free_scan_mask: |
|---|
| 1340 | | - kfree(buffer->scan_mask); |
|---|
| 1350 | + bitmap_free(buffer->scan_mask); |
|---|
| 1341 | 1351 | error_cleanup_dynamic: |
|---|
| 1342 | 1352 | iio_free_chan_devattr_list(&buffer->scan_el_dev_attr_list); |
|---|
| 1343 | | - kfree(indio_dev->buffer->buffer_group.attrs); |
|---|
| 1353 | + kfree(buffer->buffer_group.attrs); |
|---|
| 1344 | 1354 | |
|---|
| 1345 | 1355 | return ret; |
|---|
| 1346 | 1356 | } |
|---|
| 1347 | 1357 | |
|---|
| 1358 | +int iio_buffer_alloc_sysfs_and_mask(struct iio_dev *indio_dev) |
|---|
| 1359 | +{ |
|---|
| 1360 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 1361 | + const struct iio_chan_spec *channels; |
|---|
| 1362 | + int i; |
|---|
| 1363 | + |
|---|
| 1364 | + channels = indio_dev->channels; |
|---|
| 1365 | + if (channels) { |
|---|
| 1366 | + int ml = indio_dev->masklength; |
|---|
| 1367 | + |
|---|
| 1368 | + for (i = 0; i < indio_dev->num_channels; i++) |
|---|
| 1369 | + ml = max(ml, channels[i].scan_index + 1); |
|---|
| 1370 | + indio_dev->masklength = ml; |
|---|
| 1371 | + } |
|---|
| 1372 | + |
|---|
| 1373 | + if (!buffer) |
|---|
| 1374 | + return 0; |
|---|
| 1375 | + |
|---|
| 1376 | + return __iio_buffer_alloc_sysfs_and_mask(buffer, indio_dev); |
|---|
| 1377 | +} |
|---|
| 1378 | + |
|---|
| 1379 | +static void __iio_buffer_free_sysfs_and_mask(struct iio_buffer *buffer) |
|---|
| 1380 | +{ |
|---|
| 1381 | + bitmap_free(buffer->scan_mask); |
|---|
| 1382 | + kfree(buffer->buffer_group.attrs); |
|---|
| 1383 | + kfree(buffer->scan_el_group.attrs); |
|---|
| 1384 | + iio_free_chan_devattr_list(&buffer->scan_el_dev_attr_list); |
|---|
| 1385 | +} |
|---|
| 1386 | + |
|---|
| 1348 | 1387 | void iio_buffer_free_sysfs_and_mask(struct iio_dev *indio_dev) |
|---|
| 1349 | 1388 | { |
|---|
| 1350 | | - if (!indio_dev->buffer) |
|---|
| 1389 | + struct iio_buffer *buffer = indio_dev->buffer; |
|---|
| 1390 | + |
|---|
| 1391 | + if (!buffer) |
|---|
| 1351 | 1392 | return; |
|---|
| 1352 | 1393 | |
|---|
| 1353 | | - kfree(indio_dev->buffer->scan_mask); |
|---|
| 1354 | | - kfree(indio_dev->buffer->buffer_group.attrs); |
|---|
| 1355 | | - kfree(indio_dev->buffer->scan_el_group.attrs); |
|---|
| 1356 | | - iio_free_chan_devattr_list(&indio_dev->buffer->scan_el_dev_attr_list); |
|---|
| 1394 | + __iio_buffer_free_sysfs_and_mask(buffer); |
|---|
| 1357 | 1395 | } |
|---|
| 1358 | 1396 | |
|---|
| 1359 | 1397 | /** |
|---|
| .. | .. |
|---|
| 1410 | 1448 | */ |
|---|
| 1411 | 1449 | int iio_push_to_buffers(struct iio_dev *indio_dev, const void *data) |
|---|
| 1412 | 1450 | { |
|---|
| 1451 | + struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev); |
|---|
| 1413 | 1452 | int ret; |
|---|
| 1414 | 1453 | struct iio_buffer *buf; |
|---|
| 1415 | 1454 | |
|---|
| 1416 | | - list_for_each_entry(buf, &indio_dev->buffer_list, buffer_list) { |
|---|
| 1455 | + list_for_each_entry(buf, &iio_dev_opaque->buffer_list, buffer_list) { |
|---|
| 1417 | 1456 | ret = iio_push_to_buffer(buf, data); |
|---|
| 1418 | 1457 | if (ret < 0) |
|---|
| 1419 | 1458 | return ret; |
|---|