2 * linux/drivers/char/raw.c
4 * Front-end raw character devices. These can be bound to any block
5 * devices to provide genuine Unix raw character device semantics.
7 * We reserve minor number 0 for a control interface. ioctl()s on this
8 * device are used to bind the other minor numbers to block devices.
11 #include <linux/init.h>
13 #include <linux/major.h>
14 #include <linux/blkdev.h>
15 #include <linux/module.h>
16 #include <linux/raw.h>
17 #include <linux/capability.h>
18 #include <linux/uio.h>
19 #include <linux/cdev.h>
20 #include <linux/device.h>
21 #include <linux/mutex.h>
23 #include <asm/uaccess.h>
25 struct raw_device_data {
26 struct block_device *binding;
30 static struct class *raw_class;
31 static struct raw_device_data raw_devices[MAX_RAW_MINORS];
32 static DEFINE_MUTEX(raw_mutex);
33 static const struct file_operations raw_ctl_fops; /* forward declaration */
36 * Open/close code for raw IO.
38 * We just rewrite the i_mapping for the /dev/raw/rawN file descriptor to
39 * point at the blockdev's address_space and set the file handle to use
42 * Set the device's soft blocksize to the minimum possible. This gives the
43 * finest possible alignment and has no adverse impact on performance.
45 static int raw_open(struct inode *inode, struct file *filp)
47 const int minor = iminor(inode);
48 struct block_device *bdev;
51 if (minor == 0) { /* It is the control device */
52 filp->f_op = &raw_ctl_fops;
56 mutex_lock(&raw_mutex);
59 * All we need to do on open is check that the device is bound.
61 bdev = raw_devices[minor].binding;
65 igrab(bdev->bd_inode);
66 err = blkdev_get(bdev, filp->f_mode, 0);
69 err = bd_claim(bdev, raw_open);
72 err = set_blocksize(bdev, bdev_hardsect_size(bdev));
75 filp->f_flags |= O_DIRECT;
76 filp->f_mapping = bdev->bd_inode->i_mapping;
77 if (++raw_devices[minor].inuse == 1)
78 filp->f_dentry->d_inode->i_mapping =
79 bdev->bd_inode->i_mapping;
80 filp->private_data = bdev;
81 mutex_unlock(&raw_mutex);
89 mutex_unlock(&raw_mutex);
94 * When the final fd which refers to this character-special node is closed, we
95 * make its ->mapping point back at its own i_data.
97 static int raw_release(struct inode *inode, struct file *filp)
99 const int minor= iminor(inode);
100 struct block_device *bdev;
102 mutex_lock(&raw_mutex);
103 bdev = raw_devices[minor].binding;
104 if (--raw_devices[minor].inuse == 0) {
105 /* Here inode->i_mapping == bdev->bd_inode->i_mapping */
106 inode->i_mapping = &inode->i_data;
107 inode->i_mapping->backing_dev_info = &default_backing_dev_info;
109 mutex_unlock(&raw_mutex);
117 * Forward ioctls to the underlying block device.
120 raw_ioctl(struct inode *inode, struct file *filp,
121 unsigned int command, unsigned long arg)
123 struct block_device *bdev = filp->private_data;
125 return blkdev_ioctl(bdev->bd_inode, NULL, command, arg);
128 static void bind_device(struct raw_config_request *rq)
130 device_destroy(raw_class, MKDEV(RAW_MAJOR, rq->raw_minor));
131 device_create(raw_class, NULL, MKDEV(RAW_MAJOR, rq->raw_minor),
132 "raw%d", rq->raw_minor);
136 * Deal with ioctls against the raw-device control interface, to bind
137 * and unbind other raw devices.
139 static int raw_ctl_ioctl(struct inode *inode, struct file *filp,
140 unsigned int command, unsigned long arg)
142 struct raw_config_request rq;
143 struct raw_device_data *rawdev;
150 /* First, find out which raw minor we want */
152 if (copy_from_user(&rq, (void __user *) arg, sizeof(rq))) {
157 if (rq.raw_minor < 0 || rq.raw_minor >= MAX_RAW_MINORS) {
161 rawdev = &raw_devices[rq.raw_minor];
163 if (command == RAW_SETBIND) {
167 * This is like making block devices, so demand the
170 if (!capable(CAP_SYS_ADMIN)) {
176 * For now, we don't need to check that the underlying
177 * block device is present or not: we can do that when
178 * the raw device is opened. Just check that the
179 * major/minor numbers make sense.
182 dev = MKDEV(rq.block_major, rq.block_minor);
183 if ((rq.block_major == 0 && rq.block_minor != 0) ||
184 MAJOR(dev) != rq.block_major ||
185 MINOR(dev) != rq.block_minor) {
190 mutex_lock(&raw_mutex);
192 mutex_unlock(&raw_mutex);
196 if (rawdev->binding) {
197 bdput(rawdev->binding);
198 module_put(THIS_MODULE);
200 if (rq.block_major == 0 && rq.block_minor == 0) {
202 rawdev->binding = NULL;
203 device_destroy(raw_class,
204 MKDEV(RAW_MAJOR, rq.raw_minor));
206 rawdev->binding = bdget(dev);
207 if (rawdev->binding == NULL)
210 __module_get(THIS_MODULE);
214 mutex_unlock(&raw_mutex);
216 struct block_device *bdev;
218 mutex_lock(&raw_mutex);
219 bdev = rawdev->binding;
221 rq.block_major = MAJOR(bdev->bd_dev);
222 rq.block_minor = MINOR(bdev->bd_dev);
224 rq.block_major = rq.block_minor = 0;
226 mutex_unlock(&raw_mutex);
227 if (copy_to_user((void __user *)arg, &rq, sizeof(rq))) {
241 static const struct file_operations raw_fops = {
242 .read = do_sync_read,
243 .aio_read = generic_file_aio_read,
244 .write = do_sync_write,
245 .aio_write = generic_file_aio_write_nolock,
247 .release= raw_release,
249 .owner = THIS_MODULE,
252 static const struct file_operations raw_ctl_fops = {
253 .ioctl = raw_ctl_ioctl,
255 .owner = THIS_MODULE,
258 static struct cdev raw_cdev = {
259 .kobj = {.name = "raw", },
260 .owner = THIS_MODULE,
263 static int __init raw_init(void)
265 dev_t dev = MKDEV(RAW_MAJOR, 0);
268 ret = register_chrdev_region(dev, MAX_RAW_MINORS, "raw");
272 cdev_init(&raw_cdev, &raw_fops);
273 ret = cdev_add(&raw_cdev, dev, MAX_RAW_MINORS);
275 kobject_put(&raw_cdev.kobj);
279 raw_class = class_create(THIS_MODULE, "raw");
280 if (IS_ERR(raw_class)) {
281 printk(KERN_ERR "Error creating raw class.\n");
283 ret = PTR_ERR(raw_class);
286 device_create(raw_class, NULL, MKDEV(RAW_MAJOR, 0), "rawctl");
291 unregister_chrdev_region(dev, MAX_RAW_MINORS);
296 static void __exit raw_exit(void)
298 device_destroy(raw_class, MKDEV(RAW_MAJOR, 0));
299 class_destroy(raw_class);
301 unregister_chrdev_region(MKDEV(RAW_MAJOR, 0), MAX_RAW_MINORS);
304 module_init(raw_init);
305 module_exit(raw_exit);
306 MODULE_LICENSE("GPL");